before we start this portion of the lesson:

check if you have pip installed since we are going to be installing some libraries today!!!!!! if you arnt sure if you have pip, check it by running this command:

pip

if your terminal says "command not found" or something else on linux, run this:

python3 -m ensurepip --default-pip

Overview:

Pandas is a powerful tool in Python that is used for data analysis and manipulation. In this lesson, we will explore how to use Pandas to work with datasets, analyze them, and visualize the results.

Learning Objectives:

By the end of this lesson, students should be able to:

  • Understand what Pandas is and why it is useful for data analysis
  • Load data into Pandas and create tables to store it
  • Use different functions in Pandas to manipulate data, such as filtering, sorting, and grouping
  • Visualize data using graphs and charts

Question

Who here has used numpy????

(should be all odf you because all of you have used it in this class before. )

what is pandas?

no not this

this:

  • Pandas is a Python library used for data analysis and manipulation.
  • it can handle different types of data, including CSV files and databases.
  • it also allows you to create tables to store and work with your data.
  • it has functions for filtering, sorting, and grouping data to make it easier to work with.
  • it also has tools for visualizing data with graphs and charts.
  • it is widely used in the industry for data analysis and is a valuable skill to learn.
  • companies that use Pandas include JPMorgan Chase, Google, NASA, the New York Times, and many others.

Question #2 & 3:

  • which companies use pandas?
  • what is pandas?

but why is pandas useful?

  • it can provides tools for handling and manipulating tabular data, which is a common format for storing and analyzing data.
  • it can handle different types of data, including CSV files and databases.
  • it allows you to perform tasks such as filtering, sorting, and grouping data, making it easier to analyze and work with.
  • it has functions for handling missing data and can fill in or remove missing values, which is important for accurate data analysis.
  • it also has tools for creating visualizations such as graphs and charts, making it easier to communicate insights from the data.
  • it is fast and efficient, even for large datasets, which is important for time-critical data analysis.
  • it is widely used in the industry and has a large community of users and developers, making it easy to find support and resources.

Question #4:

  • why is pandas useful?

how do i flipping use it? its so hard, my puny brain cant understand it

it is actually really simple

here is numpy doing simple math:

import pandas as pd

df = pd.read_csv('yourcsvfileidcjustpickoneidiot.csv')

print(df.head())

print("Average age:", df['Age'].mean())

females = df[df['Gender'] == 'Female']
print(females)

sorted_data = df.sort_values(by='Salary', ascending=False)
print(sorted_data)
---------------------------------------------------------------------------
FileNotFoundError                         Traceback (most recent call last)
/home/chewyboba10/vscode/sushi-burrito/_notebooks/2023-04-24-pandas.ipynb Cell 10 in <cell line: 5>()
      <a href='vscode-notebook-cell://wsl%2Bubuntu/home/chewyboba10/vscode/sushi-burrito/_notebooks/2023-04-24-pandas.ipynb#X12sdnNjb2RlLXJlbW90ZQ%3D%3D?line=0'>1</a> # dummy code, just like you.
      <a href='vscode-notebook-cell://wsl%2Bubuntu/home/chewyboba10/vscode/sushi-burrito/_notebooks/2023-04-24-pandas.ipynb#X12sdnNjb2RlLXJlbW90ZQ%3D%3D?line=2'>3</a> import pandas as pd
----> <a href='vscode-notebook-cell://wsl%2Bubuntu/home/chewyboba10/vscode/sushi-burrito/_notebooks/2023-04-24-pandas.ipynb#X12sdnNjb2RlLXJlbW90ZQ%3D%3D?line=4'>5</a> df = pd.read_csv('yourcsvfileidcjustpickoneidiot.csv')
      <a href='vscode-notebook-cell://wsl%2Bubuntu/home/chewyboba10/vscode/sushi-burrito/_notebooks/2023-04-24-pandas.ipynb#X12sdnNjb2RlLXJlbW90ZQ%3D%3D?line=6'>7</a> print(df.head())
      <a href='vscode-notebook-cell://wsl%2Bubuntu/home/chewyboba10/vscode/sushi-burrito/_notebooks/2023-04-24-pandas.ipynb#X12sdnNjb2RlLXJlbW90ZQ%3D%3D?line=8'>9</a> print("Average age:", df['Age'].mean())

File /root/anaconda3/lib/python3.9/site-packages/pandas/util/_decorators.py:311, in deprecate_nonkeyword_arguments.<locals>.decorate.<locals>.wrapper(*args, **kwargs)
    305 if len(args) > num_allow_args:
    306     warnings.warn(
    307         msg.format(arguments=arguments),
    308         FutureWarning,
    309         stacklevel=stacklevel,
    310     )
--> 311 return func(*args, **kwargs)

File /root/anaconda3/lib/python3.9/site-packages/pandas/io/parsers/readers.py:680, in read_csv(filepath_or_buffer, sep, delimiter, header, names, index_col, usecols, squeeze, prefix, mangle_dupe_cols, dtype, engine, converters, true_values, false_values, skipinitialspace, skiprows, skipfooter, nrows, na_values, keep_default_na, na_filter, verbose, skip_blank_lines, parse_dates, infer_datetime_format, keep_date_col, date_parser, dayfirst, cache_dates, iterator, chunksize, compression, thousands, decimal, lineterminator, quotechar, quoting, doublequote, escapechar, comment, encoding, encoding_errors, dialect, error_bad_lines, warn_bad_lines, on_bad_lines, delim_whitespace, low_memory, memory_map, float_precision, storage_options)
    665 kwds_defaults = _refine_defaults_read(
    666     dialect,
    667     delimiter,
   (...)
    676     defaults={"delimiter": ","},
    677 )
    678 kwds.update(kwds_defaults)
--> 680 return _read(filepath_or_buffer, kwds)

File /root/anaconda3/lib/python3.9/site-packages/pandas/io/parsers/readers.py:575, in _read(filepath_or_buffer, kwds)
    572 _validate_names(kwds.get("names", None))
    574 # Create the parser.
--> 575 parser = TextFileReader(filepath_or_buffer, **kwds)
    577 if chunksize or iterator:
    578     return parser

File /root/anaconda3/lib/python3.9/site-packages/pandas/io/parsers/readers.py:933, in TextFileReader.__init__(self, f, engine, **kwds)
    930     self.options["has_index_names"] = kwds["has_index_names"]
    932 self.handles: IOHandles | None = None
--> 933 self._engine = self._make_engine(f, self.engine)

File /root/anaconda3/lib/python3.9/site-packages/pandas/io/parsers/readers.py:1217, in TextFileReader._make_engine(self, f, engine)
   1213     mode = "rb"
   1214 # error: No overload variant of "get_handle" matches argument types
   1215 # "Union[str, PathLike[str], ReadCsvBuffer[bytes], ReadCsvBuffer[str]]"
   1216 # , "str", "bool", "Any", "Any", "Any", "Any", "Any"
-> 1217 self.handles = get_handle(  # type: ignore[call-overload]
   1218     f,
   1219     mode,
   1220     encoding=self.options.get("encoding", None),
   1221     compression=self.options.get("compression", None),
   1222     memory_map=self.options.get("memory_map", False),
   1223     is_text=is_text,
   1224     errors=self.options.get("encoding_errors", "strict"),
   1225     storage_options=self.options.get("storage_options", None),
   1226 )
   1227 assert self.handles is not None
   1228 f = self.handles.handle

File /root/anaconda3/lib/python3.9/site-packages/pandas/io/common.py:789, in get_handle(path_or_buf, mode, encoding, compression, memory_map, is_text, errors, storage_options)
    784 elif isinstance(handle, str):
    785     # Check whether the filename is to be opened in binary mode.
    786     # Binary mode does not support 'encoding' and 'newline'.
    787     if ioargs.encoding and "b" not in ioargs.mode:
    788         # Encoding
--> 789         handle = open(
    790             handle,
    791             ioargs.mode,
    792             encoding=ioargs.encoding,
    793             errors=errors,
    794             newline="",
    795         )
    796     else:
    797         # Binary mode
    798         handle = open(handle, ioargs.mode)

FileNotFoundError: [Errno 2] No such file or directory: 'yourcsvfileidcjustpickoneidiot.csv'

uh oh!!! no pandas 😢

if see this error, enter these into your terminal:

pip install wheel
pip install pandas

on stack overflow, it said pandas is disturbed through pip as a wheel. so you need that too.

link to full forum if curious: https://stackoverflow.com/questions/33481974/importerror-no-module-named-pandas

ps: do this for this to work on ur laptop:

wget https://raw.githubusercontent.com/KKcbal/amongus/master/_notebooks/files/example.csv

example code on how to load a csv into a chart

import pandas as pd

# read the CSV file
df = pd.read_csv('files/example.csv')

# print the first five rows
print(df.head())

# define a function to assign each age to an age group
def assign_age_group(age):
    if age < 30:
        return '<30'
    elif age < 40:
        return '30-40'
    elif age < 50:
        return '40-50'
    else:
        return '>50'

# apply the function to the Age column to create a new column with age groups
df['Age Group'] = df['Age'].apply(assign_age_group)

# group by age group and count the number of people in each group
age_counts = df.groupby('Age Group')['Name'].count()

# print the age group counts
print(age_counts)
           Name  Age  Gender Occupation
0      John Doe   32    Male   Engineer
1    Jane Smith   27  Female    Teacher
2  Mike Johnson   45    Male    Manager
3      Sara Lee   38  Female     Doctor
4     David Kim   23    Male    Student
Age Group
30-40    7
40-50    4
<30      7
Name: Name, dtype: int64

how to manipulate the data in pandas.

import pandas as pd

# load the csv file
df = pd.read_csv('files/example.csv')

# print the first five rows
print(df.head())

# filter the data to include only people aged 30 or older
df_filtered = df[df['Age'] >= 30]

# sort the data by age in descending order
df_sorted = df.sort_values('Age', ascending=False)

# group the data by gender and calculate the mean age for each group
age_by_gender = df.groupby('Gender')['Age'].mean()

# print the filtered data
print(df_filtered)

# print the sorted data
print(df_sorted)

# print the mean age by gender
print(age_by_gender)
           Name  Age  Gender Occupation
0      John Doe   32    Male   Engineer
1    Jane Smith   27  Female    Teacher
2  Mike Johnson   45    Male    Manager
3      Sara Lee   38  Female     Doctor
4     David Kim   23    Male    Student
                Name  Age  Gender               Occupation
0           John Doe   32    Male                 Engineer
2       Mike Johnson   45    Male                  Manager
3           Sara Lee   38  Female                   Doctor
6       Robert Green   41    Male                Architect
7        Emily Davis   35  Female        Marketing Manager
8   Carlos Hernandez   47    Male             Entrepreneur
10         Kevin Lee   31    Male               Accountant
12     Jacob Johnson   34    Male                   Lawyer
13   Maria Rodriguez   39  Female               Consultant
15    Victoria Brown   42  Female  Human Resources Manager
17        Sophie Lee   30  Female          Project Manager
                Name  Age  Gender               Occupation
8   Carlos Hernandez   47    Male             Entrepreneur
2       Mike Johnson   45    Male                  Manager
15    Victoria Brown   42  Female  Human Resources Manager
6       Robert Green   41    Male                Architect
13   Maria Rodriguez   39  Female               Consultant
3           Sara Lee   38  Female                   Doctor
7        Emily Davis   35  Female        Marketing Manager
12     Jacob Johnson   34    Male                   Lawyer
0           John Doe   32    Male                 Engineer
10         Kevin Lee   31    Male               Accountant
17        Sophie Lee   30  Female          Project Manager
5        Anna Garcia   29  Female       Software Developer
14       Mark Taylor   28    Male             Web Designer
1         Jane Smith   27  Female                  Teacher
11      Rachel Baker   26  Female               Journalist
9     Melissa Nguyen   25  Female         Graphic Designer
16        Ethan Chen   24    Male       Research Assistant
4          David Kim   23    Male                  Student
Gender
Female    32.333333
Male      33.888889
Name: Age, dtype: float64

how do i put it into a chart 😩

here is how:

import pandas as pd
import matplotlib.pyplot as plt

# read the CSV file
df = pd.read_csv('files/example.csv')

# create a bar chart of the number of people in each age group
age_groups = ['<30', '30-40', '40-50', '>50']
age_counts = pd.cut(df['Age'], bins=[0, 30, 40, 50, df['Age'].max()], labels=age_groups, include_lowest=True).value_counts()
plt.bar(age_counts.index, age_counts.values)
plt.title('Number of people in each age group')
plt.xlabel('Age group')
plt.ylabel('Number of people')
plt.show()

# create a pie chart of the gender distribution
gender_counts = df['Gender'].value_counts()
plt.pie(gender_counts.values, labels=gender_counts.index, autopct='%1.1f%%')
plt.title('Gender distribution')
plt.show()

# create a scatter plot of age vs. income
plt.scatter(df['Age'], df['Income'])
plt.title('Age vs. Income')
plt.xlabel('Age')
plt.ylabel('Income')
plt.show()
---------------------------------------------------------------------------
ValueError                                Traceback (most recent call last)
/home/chewyboba10/vscode/sushi-burrito/_notebooks/2023-04-24-pandas.ipynb Cell 18 in <cell line: 9>()
      <a href='vscode-notebook-cell://wsl%2Bubuntu/home/chewyboba10/vscode/sushi-burrito/_notebooks/2023-04-24-pandas.ipynb#X23sdnNjb2RlLXJlbW90ZQ%3D%3D?line=6'>7</a> # create a bar chart of the number of people in each age group
      <a href='vscode-notebook-cell://wsl%2Bubuntu/home/chewyboba10/vscode/sushi-burrito/_notebooks/2023-04-24-pandas.ipynb#X23sdnNjb2RlLXJlbW90ZQ%3D%3D?line=7'>8</a> age_groups = ['<30', '30-40', '40-50', '>50']
----> <a href='vscode-notebook-cell://wsl%2Bubuntu/home/chewyboba10/vscode/sushi-burrito/_notebooks/2023-04-24-pandas.ipynb#X23sdnNjb2RlLXJlbW90ZQ%3D%3D?line=8'>9</a> age_counts = pd.cut(df['Age'], bins=[0, 30, 40, 50, df['Age'].max()], labels=age_groups, include_lowest=True).value_counts()
     <a href='vscode-notebook-cell://wsl%2Bubuntu/home/chewyboba10/vscode/sushi-burrito/_notebooks/2023-04-24-pandas.ipynb#X23sdnNjb2RlLXJlbW90ZQ%3D%3D?line=9'>10</a> plt.bar(age_counts.index, age_counts.values)
     <a href='vscode-notebook-cell://wsl%2Bubuntu/home/chewyboba10/vscode/sushi-burrito/_notebooks/2023-04-24-pandas.ipynb#X23sdnNjb2RlLXJlbW90ZQ%3D%3D?line=10'>11</a> plt.title('Number of people in each age group')

File /root/anaconda3/lib/python3.9/site-packages/pandas/core/reshape/tile.py:290, in cut(x, bins, right, labels, retbins, precision, include_lowest, duplicates, ordered)
    288     # GH 26045: cast to float64 to avoid an overflow
    289     if (np.diff(bins.astype("float64")) < 0).any():
--> 290         raise ValueError("bins must increase monotonically.")
    292 fac, bins = _bins_to_cuts(
    293     x,
    294     bins,
   (...)
    301     ordered=ordered,
    302 )
    304 return _postprocess_for_cut(fac, bins, retbins, dtype, original)

ValueError: bins must increase monotonically.

uh oh!!!! another error!??!!??!?! install this library:

pip install matplotlib
import pandas as pd
import matplotlib.pyplot as plt
import numpy as np

# read the CSV file
df = pd.read_csv('files/example.csv')

# define age groups
age_groups = ['<30', '30-40', '40-50', '>50']

# create a new column with the age group for each person
df['Age Group'] = pd.cut(df['Age'], bins=[0, 30, 40, 50, np.inf], labels=age_groups, include_lowest=True)

# group by age group and count the number of people in each group
age_counts = df.groupby('Age Group')['Name'].count()

# create a bar chart of the age counts
age_counts.plot(kind='bar')

# set the title and axis labels
plt.title('Number of People in Each Age Group')
plt.xlabel('Age Group')
plt.ylabel('Number of People')

# show the chart
plt.show()

magic!!!!!!

Hacks

  1. make your own data using your brian, google or chatgpt, should look different than mine.
  2. modify my code or write your own
  3. output your data other than a bar graph.
  4. write an 850+ word essay on how pandas, python or irl, affected your life. If AI score below 85%, then -1 grading point
  5. answer the questions below, the more explained the better.

Questions

  1. What are the two primary data structures in pandas and how do they differ? The two primary data structures in pandas are Series and DataFrame. A Series is a one-dimensional array-like object that can hold any data type. A DataFrame is a two-dimensional table-like data structure with rows and columns, similar to a spreadsheet.
  2. How do you read a CSV file into a pandas DataFrame? To read a CSV file into a pandas DataFrame, you can use the read_csv function in pandas.
  3. How do you select a single column from a pandas DataFrame? To select a single column from a pandas DataFrame, you can use the indexing operator [] with the column name.
  4. How do you filter rows in a pandas DataFrame based on a condition? To filter rows in a pandas DataFrame based on a condition, you can use boolean indexing.
  5. How do you group rows in a pandas DataFrame by a particular column? To group rows in a pandas DataFrame by a particular column, you can use the groupby method.
  6. How do you aggregate data in a pandas DataFrame using functions like sum and mean? To aggregate data in a pandas DataFrame using functions like sum and mean, you can use the agg method.
  7. How do you handle missing values in a pandas DataFrame? To handle missing values in a pandas DataFrame, you can use the fillna method to fill in missing values with a specific value or method, or you can use the dropna method to remove rows with missing values.
  8. How do you merge two pandas DataFrames together? To merge two pandas DataFrames together, you can use the merge method.
  9. How do you export a pandas DataFrame to a CSV file? To export a pandas DataFrame to a CSV file, you can use the to_csv method.
  10. What is the difference between a Series and a DataFrame in Pandas? The main difference between a Series and a DataFrame in pandas is that a Series is a one-dimensional array-like object, while a DataFrame is a two-dimensional table-like data structure. A Series can be thought of as a single column of a DataFrame, while a DataFrame can have multiple columns.

note

all hacks due saturday night, the more earlier you get them in the higher score you will get. if you miss the due date, you will get a 0. there will be no tolerance.

no questions answered

Tonight- 2.9

Friday Night- 2.8

Saturday Night - 2.7

Sunday Night - 0.0

questions answered

Tonight- 3.0

Friday Night- 2.9

Saturday Night - 2.8

Sunday Night - 0.0

wdfasdf

import pandas as pd
import matplotlib.pyplot as plt

# Read the CSV file
df = pd.read_csv('files/data.csv')

# Create a scatter plot
plt.scatter(df['Duration'], df['Calories'], color='blue')

# Set the axis labels and title
plt.xlabel('Duration')
plt.ylabel('Calories')
plt.title('Calories vs Duration')

# Display the plot
plt.show()

Numpy

from skimage import io
photo = io.imread('images/waldo.jpg')
type(photo)
import matplotlib.pyplot as plt
plt.imshow(photo)
photo.shape
(461, 700, 3)
plt.imshow(photo[210:350, 425:500])
<matplotlib.image.AxesImage at 0x7f2ca2a7a1c0>
import numpy as np
import matplotlib.pyplot as plt

# Create a 2D array representing the complex plane
x_min, x_max = -2.5, 1.5
y_min, y_max = -2, 2
resolution = 1000
x, y = np.meshgrid(np.linspace(x_min, x_max, resolution), np.linspace(y_min, y_max, resolution))
c = x + 1j*y

# Create a 2D array representing the iterations required for each point to escape the Mandelbrot set
z = np.zeros_like(c)
for i in range(100):
    z = z**2 + c
    mask = np.abs(z) > 2
    z[mask] = 0
    if not np.any(mask):
        break

# Create a grayscale image of the Mandelbrot set
plt.imshow(i - np.log2(np.log2(np.abs(z))) if np.any(mask) else 0, cmap='gray')
plt.axis('off')
plt.show()
/tmp/ipykernel_10310/3948469102.py:21: RuntimeWarning: divide by zero encountered in log2
  plt.imshow(i - np.log2(np.log2(np.abs(z))) if np.any(mask) else 0, cmap='gray')
/tmp/ipykernel_10310/3948469102.py:21: RuntimeWarning: invalid value encountered in log2
  plt.imshow(i - np.log2(np.log2(np.abs(z))) if np.any(mask) else 0, cmap='gray')
import numpy as np
import matplotlib.pyplot as plt

# Generate some random data
data = np.random.rand(10, 10)

# Create a heatmap using imshow
plt.imshow(data, cmap='coolwarm')

# Add a colorbar for reference
plt.colorbar()

# Show the plot
plt.show()

Data/Predictive Analysis

How can Numpy and Pandas be utilized for data preprocessing in predictive analysis? Both Numpy and Pandas are tools that can be used to clean, standardize, and transform data to prepare it for predictive analysis. While Numpy is preferred for numerical data due to its ability to perform mathematical operations, Pandas offers a wider range of features to handle diverse data types.

What are some machine learning algorithms that can be used for predictive analysis, and how do they differ? There are several algorithms that can be used for predictive analysis, including linear regression, decision trees, random forests, neural networks, and support vector machines. These algorithms differ in terms of their specific applications, such as forecasting continuous outcomes or identifying optimal boundaries between different classes within a dataset.

What are some real-world applications of predictive analysis across different industries? Predictive analysis has various applications in different industries, including predicting weather forecasts, determining the likelihood of medical conditions based on scans, predicting sports game outcomes, and forecasting stock market trends.

What is the role of feature engineering in predictive analysis, and how does it impact model accuracy? Feature engineering involves selecting and manipulating variables to create a predictive analysis model. This technique can improve model accuracy by identifying crucial information that highlights patterns within the data.

How can machine learning models be deployed in real-time applications for predictive analysis? Real-time applications can use machine learning models to make predictions based on user interactions. For instance, TikTok can use machine learning models to analyze user preferences and suggest videos that align with their interests.

What are the limitations of Numpy and Pandas, and when might other data analysis tools be necessary? While Numpy is memory-efficient and preferred for numerical data, Pandas utilizes more memory but can handle diverse data types. For extremely large datasets or complex syntax, alternative tools may be necessary. Additionally, Pandas has a steep learning curve and limited documentation, and it is not compatible with 3D matrices.

How can predictive analysis improve decision-making and optimize business processes? Predictive analysis can facilitate better decision-making and optimize business processes by enabling businesses to forecast probable outcomes and tailor customer experiences to increase profits.