Where communities thrive


  • Join over 1.5M+ people
  • Join over 100K+ communities
  • Free without limits
  • Create your own community
People
Repo info
Activity
  • Aug 01 18:05
    mr-devs commented #403
  • Jul 30 19:12

    edsu on v2.3.12

    (compare)

  • Jul 30 19:12

    edsu on main

    do not set start-time when unti… (compare)

  • Jul 30 17:34

    edsu on v2.3.11

    (compare)

  • Jul 30 17:34

    edsu on main

    new version (compare)

  • Jul 30 17:25

    edsu on main

    black (compare)

  • Jul 30 15:12

    edsu on main

    Check usernames/ids in timeline… (compare)

  • Jul 30 15:09
    edsu edited #514
  • Jul 30 15:08
    edsu edited #514
  • Jul 30 15:08
    edsu commented #514
  • Jul 30 14:15
    igorbrigadir commented #514
  • Jul 30 14:14
    SamHames commented #514
  • Jul 30 13:53
    edsu opened #514
  • Jul 30 13:48
    edsu commented #513
  • Jul 30 13:48

    edsu on main

    add progress bar to counts comm… Merge pull request #513 from Do… (compare)

  • Jul 30 13:48
    edsu closed #513
  • Jul 30 12:40
    igorbrigadir opened #513
  • Jul 30 12:38

    igorbrigadir on counts-progress

    add progress bar to counts comm… (compare)

  • Jul 29 15:29
    igorbrigadir commented #512
  • Jul 29 13:48
    edsu edited #512
Ed Summers
@edsu
search_archive.py doesn't work yet
the idea would be you could run it with a query and a directory where the twitter result files would live
search_archive.py --search ferguson --archive_dir=/mnt/tweets/ferguson
Renato Gabriele
@remagio
my fault, haha
Ed Summers
@edsu
the first time you run it, it would get as many tweets as it can and write them to /mnt/tweets/ferguson/tweets-0001.json
Renato Gabriele
@remagio
miscmatch between our discussion and source ;-)
Ed Summers
@edsu
the next time you run it it will look at the first tweet in tweets-00001.json and use that twitter id as the minimum id to archive, and write the tweets as /mnt/tweets/ferguson/tweets-0002.json
that's what i was thinking anyway
i didn't like the way the old twarc named files
Renato Gabriele
@remagio
right
I think it works and ppl will like this
Ed Summers
@edsu
ok let me see if i can get it working
maybe you can try it out :-)
Renato Gabriele
@remagio
:-)
Anyway, THX!
You contributions in this kind of stuff is great
Have a good day @edsu
Ed Summers
@edsu
hopefully i'll have something for you to try in an hour
thanks for working w/ me on it!
Renato Gabriele
@remagio
You are welcome
We could discuss about nosql integration soon if you want
Ed Summers
@edsu
super to have some in italy trying it out btw -- makes a big difference to have it not just be a US thing
sure
some -> someome
someone!
Renato Gabriele
@remagio
I see, but we have ridiculous "numbers" comparing with US
I mean, media say "Incredible, Italian twitter scene crazy for next Presidente"
But it's less than 200k tweets most bot or smm with botnets.
Ed Summers
@edsu
:-) yeah, but i think it's interesting how social media is being used in different cultures/contexts
Renato Gabriele
@remagio
Sure, here the hype is very high only because we have some journalists that are at the same time: journalist+SMM+Political consultants. And are leader in TV&Papers. So every media say everyday: People are creasy on Twitter for blah blah blah
And usual are only 30k-50k tweets
But got prime rate news on TV.
Ed Summers
@edsu
i see yeah
Renato Gabriele
@remagio
Here comes twarc
;-)
plus other tools.
I lived in a few country but I think Italy, by example, is still in the early phase of analysis of socials.
I mean It's still used to support outside twitter points.
Ed Summers
@edsu
i think we all are still figuring out what social media means ; particularly in the political arena
Renato Gabriele
@remagio
yes, here I think there will be lots to learn b/c the entire smm community is involved in.
politics.
But happen something not possible in US.
The same persons create opinions using CnC plus consultants plus volunteer.
Ed Summers
@edsu
quick question: would it be ok if this search_archive.py wrote the data as gzip files?
or would it be easier if they are uncompressed to start with?
Renato Gabriele
@remagio
uncompressed better
Ed Summers
@edsu
ok
Renato Gabriele
@remagio
so using pipe it can be checked
Ed Summers
@edsu
zcat is nice on linux
zcat file.gz | grep foo
Renato Gabriele
@remagio
not yes used in bash scripts