It’s a way to take a complete Facebook archive, ingest it, and import it into your website in a way that is properly formatted and allows for others to reply, like, ETC. each piece of content.
It’s taken me a year to get it working, and it’s still a bit of a heavy lift for people who aren’t tech heads. Trying to make that easier so other people can do it.
You need to be comfortable with the command line and http status codes at the moment, but I’m trying to make this a little easier for other users because it’s still a bit of a lift.
There are still a few things it doesn’t do, like resume ingesting if for some reason the process is interrupted, and I need to research how to tell python scripts how to look for a range of status codes instead of just one.
@bekopharm I’d be more than happy to try to help get things working for you. Let me read through your post real quick, and if you want we can schedule some time next week in the evening and I can either walk you through it, or I can do the import for you. That would take a few days since I’d need the archive and right now anything that does not have a photo attached has to go through a yes/no process, but I’m sure we can get it going.
@bekopharm OK just looked at your post and will make some other modifications to the scripts to accept the new data, since it’s just looping through anything with the post field and looking for a filename.
@bekopharm I’m going to upload the modified version but I think I need to wait until I finish this archive import since I already have a copy of Ditchbook in my home directory. Plus I’ve found some other mods I need to make because if it times out it won’t retry the connection and won’t resume, so on big archives you have to start the whole process over again. I think I’m going to make a couple of copies though, one for automatic upload of everything provided you can get around time-outs and one for paging through. Trying to make it as easy as possible for people to run this. And I’m still testing the changes with the new archive file. Need to download one that matches the format mentioned in your issue but still waiting on the redownload of mine to see if the current format matches that.
@bekopharm Can you do me a favor and drop me your archive? I’m not going to read through it or anything but I need to see exactly how each post is organized/what fields FB assigns where.
@arush Uploading it to my online data dump but how to contact you for the URL? I can’t find intersecting rel=”me” links on your website that would allow me to message you. Otherwhise feel free to poke around all the way you like in this. The really important posts (for me) I did syndicate back manually already months ago.
Is this similar to your tweeting from WP?
It’s a way to take a complete Facebook archive, ingest it, and import it into your website in a way that is properly formatted and allows for others to reply, like, ETC. each piece of content.
It’s taken me a year to get it working, and it’s still a bit of a heavy lift for people who aren’t tech heads. Trying to make that easier so other people can do it.
Nice! Sounds like this could be useful. Let me know if and or when you are cool to have others kicking the tires too.
ditch book? SOunds intresting. HOw do I use this?
You need to be comfortable with the command line and http status codes at the moment, but I’m trying to make this a little easier for other users because it’s still a bit of a lift.
There are still a few things it doesn’t do, like resume ingesting if for some reason the process is interrupted, and I need to research how to tell python scripts how to look for a range of status codes instead of just one.
Right now if it doesn’t get the exact status code it’s looking for it’ll bail, and you have to start the publishing process from the beginning.
@arush Highly interested. It never worked for me: beko.famkos.net/2019/08/3…
github.com/cleverdev…
@bekopharm I’d be more than happy to try to help get things working for you. Let me read through your post real quick, and if you want we can schedule some time next week in the evening and I can either walk you through it, or I can do the import for you. That would take a few days since I’d need the archive and right now anything that does not have a photo attached has to go through a yes/no process, but I’m sure we can get it going.
@bekopharm OK just looked at your post and will make some other modifications to the scripts to accept the new data, since it’s just looping through anything with the post field and looking for a filename.
@arush Hm? Are you not uploading your modified import to Github or something?
I mean thanks but there is really no need that you make more work of this as really needed for yourself.
@bekopharm I’m going to upload the modified version but I think I need to wait until I finish this archive import since I already have a copy of Ditchbook in my home directory. Plus I’ve found some other mods I need to make because if it times out it won’t retry the connection and won’t resume, so on big archives you have to start the whole process over again. I think I’m going to make a couple of copies though, one for automatic upload of everything provided you can get around time-outs and one for paging through. Trying to make it as easy as possible for people to run this. And I’m still testing the changes with the new archive file. Need to download one that matches the format mentioned in your issue but still waiting on the redownload of mine to see if the current format matches that.
@bekopharm Can you do me a favor and drop me your archive? I’m not going to read through it or anything but I need to see exactly how each post is organized/what fields FB assigns where.
@arush Uploading it to my online data dump but how to contact you for the URL? I can’t find intersecting rel=”me” links on your website that would allow me to message you. Otherwhise feel free to poke around all the way you like in this. The really important posts (for me) I did syndicate back manually already months ago.
@bekopharm Thanks for letting me know re: the re=”me” links on my site I’ve been meaning to fix that. Email is amanda@customerservant.com
@arush rel=”me”