Engine of forum

Neia Neutuladh neia at ikeran.org
Wed Aug 22 16:06:38 UTC 2018


On Wednesday, 22 August 2018 at 15:17:36 UTC, Jesse Phillips 
wrote:
> It is weird that you make loosing current and historical pull 
> requests is minor

It would be disruptive. However, work could resume rather quickly.

The disruption would be reduced if we had a periodic job set up 
to mirror github pull requests. There is at least one export 
tool, but it seems to get only the titles and not the comments or 
patches.

> compared to:
>
> * Having all the data readily available for search engines to 
> have archived (today, not tomorrow).
> * Having an established forum/newsgroup readily available to 
> handle the load of new questions.
>
> I just don't see data retention and recovery for StackOverflow 
> to be a concern for making such a choice. Even if it did take 
> weeks or months to host the historical data, risk should be 
> weighed against possible benefit from visibility and growth 
> from heavily using StackOverflow.

And similarly, the choice of Github instead of a self-hosted 
system is weighed against requiring people to sign up with a 
private gitlab instance. Also similarly, the disruption would be 
reduced if we had a periodic job set up to handle long-term 
stackoverflow unavailability in advance.

I'm a little paranoid about centralized services like Github. I'd 
prefer a federated service for source control / project 
management, where you could easily fork projects from my server 
to yours and send back pull requests. Then there would be no 
extra cost for hosting your own vs using an existing instance.

I've been low-key thinking about making a federated github, one 
where exporting your data is as simple as a `git clone; git 
submodule update --init`. Probably nothing will come of it, 
though.


More information about the Digitalmars-d mailing list