Skip to content

Olatokunbo360/We_Rate_Dogs

Repository files navigation

Data Analysis on X (Formerly Twitter) We_Rate_Dogs Account

Overview

This repository contains a data analysis project focused on gathering, assessing, cleaning, storing, wrangling, analysing and visualising a dataset. The dataset is the tweet archive of X (formerly Twitter) user @dog_rates, also known as WeRateDogs. WeRateDogs is an X account that rates people's dogs with a humorous comment about the dog. These ratings almost always have a denominator of 10. The numerators, though? Almost always greater than 10. 11/10, 12/10, 13/10, etc. which indicates how high the dog is rated. WeRateDogs has over 4 million followers and has received international media coverage.

Dataset

Three datasets would be worked on in this project

Datasets and Files

twitter-archive-enhanced-2.csv: This contains a fitered down list of tweets from the WeRateDogs X archive that have ratings. Some tweets do not have ratings. The file would still need to be checked to ensure that it was cleaned properly.

Data via X API: This is data gathered through the use of the X API. Access to this API is requested from X. You could use the directly downloaded file (twitter-archive-enhanced-2.csv) or you could use this API and download the required file from X directly after you have requested and been granted access by X.

image-predictions.tsv: This is a file created from a neural link (see the AIPND-Pre-Trained repository) that can classify dog breeds.

wrangle_act.ipynb: Jupyter notebook that contains the Python code for data analysis

wrangle_report.pdf:PDF file that contains a report on how the data was wrangled and analysed

act_report.pdf: PDF file that contains a report on the visualisations and insights gotten from the data analysis exercise.

tweet-json.txt: txt file that contains all the saved tweets from the X API returned as JSON objects.

README.md: This file, providing an overview of the project and instructions for use.

requirements.txt: This file provides the list of the Python libraries used for this project.

Dependencies

The analysis is conducted using Python programming language and several libraries including:

Pandas: For data manipulation and analysis.

Matplotlib: For data visualization.

Seaborn: For statistical data visualization.

NumPy: For numerical computing.

Tweepy: For accessing X API

Requests: For sending HTTP requests

Json: For working with JSON data

timeit: For determining the time for the code execution

os: For creating and accessing files

Usage

To run the analysis on your local machine, follow these steps:

Clone this repository to your local machine using the following command:

bash

Copy code git clone https://github.com/Olatokunbo360/We_Rate_Dogs.git

Navigate to the project directory:

bash

Copy code cd We_Rate_Dogs

Install the required dependencies. You can use the following command to install dependencies using pip:

Copy code pip install -r requirements.txt

Once the dependencies are installed, you can open the Jupyter Notebook wrangle_act.ipynb to view the analysis and execute the code cells.

Follow the instructions provided in the notebook to explore the dataset, analyze trends, and visualize insights related to the dataset.

Contributing

Contributions to this project are welcome. If you have suggestions for improvements or new analyses, please feel free to open an issue or submit a pull request.

License This project is licensed under the MIT License - see the LICENSE file for details.

By [Yusuf Sanni] - [yusufsanni2003@yahoo.co.uk]

About

Udacity Project for Data Analysis Nanodegree - We Rate Dogs Dataset

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors