by

Where communities thrive


  • Join over 1.5M+ people
  • Join over 100K+ communities
  • Free without limits
  • Create your own community
People
Repo info
Activity
  • Jul 07 11:02
    tddschn commented #338
  • Jul 07 10:51
    edsu commented #338
  • Jul 07 07:15
    tddschn opened #338
  • Jul 06 20:01

    edsu on v1.10.1

    (compare)

  • Jul 06 20:00

    edsu on main

    allow access key/secret to be e… Merge branch 'main' of github.c… (compare)

  • Jul 06 09:41
    szeyd commented #259
  • Jul 02 11:40

    edsu on main

    Fix broken links: master->main Merge pull request #337 from hu… (compare)

  • Jul 02 11:40
    edsu closed #337
  • Jul 02 11:06
    hugovk opened #337
  • Jul 01 01:06
    edsu commented #336
  • Jul 01 01:05
    edsu commented #336
  • Jul 01 01:04
    mspanish closed #336
  • Jul 01 01:04
    mspanish commented #336
  • Jul 01 00:51
    edsu commented #336
  • Jul 01 00:50
    edsu commented #336
  • Jul 01 00:49
    edsu commented #336
  • Jul 01 00:46
    mspanish commented #335
  • Jul 01 00:22
    mspanish commented #335
  • Jul 01 00:01
    mspanish opened #336
  • Jun 30 23:29
    mspanish commented #335
Ed Summers
@edsu
I'm sorry I do not understand.
Renato Gabriele
@remagio
Checking search_archive.py I realize it works on save files by twarc
saved
do I miss something?
Ed Summers
@edsu
search_archive.py doesn't work yet
the idea would be you could run it with a query and a directory where the twitter result files would live
search_archive.py --search ferguson --archive_dir=/mnt/tweets/ferguson
Renato Gabriele
@remagio
my fault, haha
Ed Summers
@edsu
the first time you run it, it would get as many tweets as it can and write them to /mnt/tweets/ferguson/tweets-0001.json
Renato Gabriele
@remagio
miscmatch between our discussion and source ;-)
Ed Summers
@edsu
the next time you run it it will look at the first tweet in tweets-00001.json and use that twitter id as the minimum id to archive, and write the tweets as /mnt/tweets/ferguson/tweets-0002.json
that's what i was thinking anyway
i didn't like the way the old twarc named files
Renato Gabriele
@remagio
right
I think it works and ppl will like this
Ed Summers
@edsu
ok let me see if i can get it working
maybe you can try it out :-)
Renato Gabriele
@remagio
:-)
Anyway, THX!
You contributions in this kind of stuff is great
Have a good day @edsu
Ed Summers
@edsu
hopefully i'll have something for you to try in an hour
thanks for working w/ me on it!
Renato Gabriele
@remagio
You are welcome
We could discuss about nosql integration soon if you want
Ed Summers
@edsu
super to have some in italy trying it out btw -- makes a big difference to have it not just be a US thing
sure
some -> someome
someone!
Renato Gabriele
@remagio
I see, but we have ridiculous "numbers" comparing with US
I mean, media say "Incredible, Italian twitter scene crazy for next Presidente"
But it's less than 200k tweets most bot or smm with botnets.
Ed Summers
@edsu
:-) yeah, but i think it's interesting how social media is being used in different cultures/contexts
Renato Gabriele
@remagio
Sure, here the hype is very high only because we have some journalists that are at the same time: journalist+SMM+Political consultants. And are leader in TV&Papers. So every media say everyday: People are creasy on Twitter for blah blah blah
And usual are only 30k-50k tweets
But got prime rate news on TV.
Ed Summers
@edsu
i see yeah
Renato Gabriele
@remagio
Here comes twarc
;-)
plus other tools.
I lived in a few country but I think Italy, by example, is still in the early phase of analysis of socials.
I mean It's still used to support outside twitter points.
Ed Summers
@edsu
i think we all are still figuring out what social media means ; particularly in the political arena
Renato Gabriele
@remagio
yes, here I think there will be lots to learn b/c the entire smm community is involved in.
politics.
But happen something not possible in US.
The same persons create opinions using CnC plus consultants plus volunteer.
Ed Summers
@edsu
quick question: would it be ok if this search_archive.py wrote the data as gzip files?
or would it be easier if they are uncompressed to start with?
Renato Gabriele
@remagio
uncompressed better
Ed Summers
@edsu
ok