21.01.2015 Views

color version - PET: Python Entre Todos - Python Argentina

color version - PET: Python Entre Todos - Python Argentina

color version - PET: Python Entre Todos - Python Argentina

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

(presidential advisors for example).<br />

In this case the value of the information would be 10,000, and since we define<br />

value as “much” and what value we define as “little”, this value is “large” (much).<br />

This, we can consider as correct as any medium that can influence important<br />

people should have a high value.<br />

Keeping the values, but changing our function by AUDIENCE + IMPACT the<br />

value of the information would be 1010 which, keeping the above reasoning<br />

remains high.<br />

Now, if we replace the value of the audience by a large number 1000, and keep the<br />

impact, the value of the original function would be 1 million and the second case<br />

in 2000.<br />

If we consider we now likely impact on 1000 presidential advisors 2000 value<br />

remains small, since we are presence of a medium likely to generate a global<br />

impact. Which shows that multiplication represents much better variation of<br />

parameters.<br />

• The value vanishes in the absence of audience or impact: When the audience<br />

or impact are 0 (zero) (No one sees or pays attention to the medium) the value of<br />

information is also 0 (zero).<br />

This is not trivial because it suggests that the information is worthless if nobody<br />

cares to see it or no one pays attention.<br />

Infopython<br />

Since there is a wide variety of public services that extract statistics and data on new<br />

media (web, twitter, etc), including:<br />

• Klout (http://klout.com/)<br />

• Compete (http://www.compete.com/)<br />

• Alexa (http://www.alexa.com/)<br />

To quote a few. Infopython focuses on providing a simple API to value through<br />

agenda-setting (in the future to implement other theories) to the media regardless of<br />

type, using aforementioned services<br />

Architecture:<br />

Analyzing every layer:<br />

• Internet Service: Corresponds to the different services that exist<br />

on the Web for statistics and data mining of new media.<br />

• Other Sources: They are other data which feeds Infopython, such as databases,<br />

excel templates, etc.<br />

• Scipy: It is an open source library of algorithms and<br />

mathematical tools.<br />

This handles the necessary number crunching.<br />

• Third Parties Apis: These are third-party libraries that connect to services that<br />

exist in the network. For example:<br />

• Tweepy used to manipulate data from twitter.<br />

• Koutpy that connects to Klout<br />

• Session: This sub-layer is a module that is responsible for centralizing all<br />

necessary settings to access internet services.<br />

• Interpolation Normalization: This is a layer of abstraction for different<br />

interpolators which has scipy and defines some new all with the same API.<br />

• API Normalization: will convert all the answers of all Internet services and third<br />

party APIs to common structures (dictionaries) using if needed information<br />

contained in the session.<br />

• Information Sources: These are the classes that represent our sources. These are<br />

connected in a “auto-magical” way to the various standardized API’s.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!