Found 237 repositories(showing 30)
amankaushik
The information system chosen for the project was a stock investment management website providing live prices, historical data, news articles, etc and also basic analysis and recommendations using data mining techniques. 1. Crawling and parsing Yahoo-Finance, Reuters and Twitter data (Java, twitter4j). 2. Web Interface using J2EE and Struts-2 framework. jQuery (highstocks lib) for showing technical charts. 3. Database integration, data cleaning, feature selection on the collected data and applying linear regression and classification algorithms : SVM, Naive Bayes to produce detailed analysis and recommendations.
kapilkchaurasia
It contain various script on web crawling/ data mining of social web(RSS,facebook,twitter,Linkedin)
mpfarmer
Tweets, Twitter, Search, Crawling, Tool, Big Data Analysis, Data Mining, Time Series Analysis
phamphihungbk
🤖 Small python scripts to crawl data from FIT (Facebook, Instagram, Twitter)
alpengeist
Crawls Twitter friends and gets geolocation data from Yahoo Places; stores graph in neo4j
syenirasheila
This project aims to analyze the sentiment about IKN (Ibu Kota Negara), the New National Capital in Indonesia. Data is obtained through crawling Twitter data related to IKN discussion topics. Furthermore, the data is analyzed using the SVM classification method by combining it with the Query Expansion technique to produce better model performance.
utadstriker9
Crawling Data Twitter Menggunakan Tweepy Python dan Melakukan Analisis Sentimen Berdasarkan Polarity Text.
FathonyS073
Crawling Twitter dengan Tweet-Harvest Twitter Crawler with Tweet-Harvest adalah sebuah proyek untuk melakukan crawling tweet menggunakan Tweet-Harvest, sebuah tool berbasis SNScrape untuk mengumpulkan data dari Twitter (X) tanpa memerlukan API.
MohammadNuramin
spamming. The presence of spam on web services such as search engines, email providers or online networking services can be manifested in many ways including spam advertising, malicious links, fake news or fake friends but also manipulation attempts. For online social network, tracking and controlling spammers are of the upmost importance due to the security risk but also for the credibility of the information that they disseminate. The objective of this project is to study Twitter’s social spam by means of both data mining, machine learning and data analysis techniques (that you have learned so far in any course!) using a dataset containing information on 767 social spammers and legitimate users crawled from Twitter in November and December 2014 and July 2018. In this project you have firstly to solve the spam detection problem and secondly to analyse the dataset using methods presented during the lessons of data mining.
andrebudiman
Data hasil Crawling pada Twitter melalui Twitter API
joshuagohez
a generative ai application that crawls Linkedin & Twitter data about a person and customises a starting message with them via a LLM
efimeida
Source Code Python untuk Crawling Data Twitter
yahdiindrawan
Crawling tweet data about Covid-19 in Indonesian from Twitter API for sentiment analysis into 3 categories, positive, negative and neutral
alichoumane
This platform offers a GUI to help crawling Twitter data (graphs, tweets, full public profiles) for research purposes. It is built on the top of the Twitter4J library.
mengfeidu
This is a project of DATA130007.01 Social Network Mining(2021). We crawl Twitter data from 2020 us presidential election and try to analyze the effect of bots during this election.
FachriezalNugraha
Crawling data Twitter dengan mengggunakan JupyterNotebook dan Library tweepy
ChangUk
Crawling machine for gathering Twitter data (written in Java)
Derida23
Script example for crawling data in twitter, to analyze sentiment from internet focused on twitter.
Cara crawling data dengan menggunakan twitter
matteoredaelli
dump_tweets.R is a tool for searching tweets and crawl (recursively) users from twitter. Data are then saved to a MySQL database and can finally be exported to .RData files
dani-amirtharaj
Aggregated data from Twitter, New York Times and Common Crawl, applied a big data analytic method (Hadoop MapReduce) on data obtained, and built a data visualization product with Tableu.
JulianLopezB
Social network analysis upon twits containing hashtags with political content during campaigns for general elections to be held in Argentina. Data was crawled from Twitter API in 09/30/2019, tracking the keywords '#SiSePuede' and '#YoVoy'.
socket-var
Perform big data analysis on New york times, Twitter and Common Crawl APIs
pramudianzaheka92
Repository ini merupakan sebuah projek scraping & crawling data dari 3 topik yaitu pandemi, politik dan ekonomi. Sumber scraping & crawling yang digunakan adalah melalui media sosial (twitter) dan website (TurnBackHoax, Detik.com dan MetroTVNews.com)
sebenns
A project for crawling accounts via Twitter API, classifying and analyzing their contents via Google Natural Language AI and importing resulting classified data into a neo4j graph database.
fulcrum6378
A web-based application which crawls profiles on Twitter for all of their tweets, all tweets related to them, including their attachments, statistics and data of their authors. Main data is stored in an SQLite database and all media are downloaded. Then it'll be able to reconstruct a Twitter profile in front-end.
mikel-code
Crawler for Twitter Search that don't use Twitter's APIs. The crawled data is not as clean as the one obtained by the APIs, but the benefits are you can get rid of the API's rate limits and restrictions. Based on https://github.com/jonbakerfish/TweetScraper.
• Aggregated data from following sources suing the respective application programing interface(API). 1. NY Times 2. Twitter 3. Common Crawl - Applied classical big data analytic method of Map Reduce to the unstructured data collected by the above mentioned sources. - Stored the data collected on WORM infrastructure Amazon Web Services (AWS) S3 AND EC2. - Visualized using Tableau to deduce trends in the processed Data
Collected data about from three sources, one opinion-based social media in twitter, research data in New York Times, and the third is the common crawl data for the same topic or key phrase, and from similar time periods. Processed the three data sets collected individually using classical big data methods like Map Reduce in Google Dataproc Clusters. And then compared the outcomes using popular visualization methods in tableau.
AndySF002
example Code for Crawling data