from dsc80_utils import *
Welcome to DSC 80! 🎉
Agenda¶
- Who are we?
- What does a data scientist do?
- What is this course about, and how will it run?
- The data science lifecycle.
- Example: What's in a name?
Instructor: Samuel Lau (call me Sam)¶
Prof. Sam Lau¶
- Assistant Teaching Professor, HDSI, UCSD
- Personal: https://www.samlau.me/
I design curriculum and invent tools for teaching programming and data science.
Bio: Ph.D. UCSD (2023), M.S. UC Berkeley (2018), B.S. UC Berkeley (2017).
- My first year as a professor at UCSD 🎉
- Pandas Tutor visualizes
pandas
code 📊: https://pandastutor.com/ - Learning Data Science, a free textbook on data science 📚: https://learningds.org
- Outside the classroom 👨🏫: cooking, eating out, woodworking
Course staff¶
In addition to the instructor, we have several staff members who are here to help you in discussion, office hours, and on Ed:
- 1 graduate TA: Praveen Nair.
- 11 undergraduate tutors: Andrew Yang, Aritra Das, Diego Zavalza, Gabriel Cha, Guoxuan (Jason) Xu, Jasmine Lo, Luran Zhang, Mizuho Fukuda, Qirui (Sara) Zheng, Sunan Xu, and Ylesia Wu.
Learn more about them at dsc80.com/staff.
What is data science? 🤔¶
The DSC 10 approach¶
In DSC 10, we told you that data science is about drawing useful conclusions from data using computation. In DSC 10, you:
- Used Python to explore and visualize data.
- Used simulation to make inferences about a population, given just a sample.
- Made predictions about the future given data from the past.
Let's look at a few more definitions of data science.
What is data science?¶
There isn't agreement on which "Venn Diagram" is correct!
- Why not? The field is new and rapidly developing.
- Make sure you're solid on the fundamentals, then find a niche that you enjoy.
- Read Taylor, Battle of the Data Science Venn Diagrams.
What does a data scientist do?¶
The chart below is taken from the 2016 Data Science Salary Survey, administered by O'Reilly. They asked respondents what they spend their time doing on a daily basis. What do you notice?
The chart below is taken from the followup 2021 Data/AI Salary Survey, also administered by O'Reilly. They asked respondents:
What technologies will have the biggest effect on compensation in the coming year?
What does a data scientist do?¶
My take: in DSC 80, and in the DSC major more broadly, we are training you to ask and answer questions using data.
As you take more courses, we're training you to answer questions whose answers are ambiguous – this uncertainly is what makes data science challenging!
Let's look at some examples of data science in practice.
An excerpt from the article:
Global warming is precisely the kind of threat humans are awful at dealing with: a problem with enormous consequences over the long term, but little that is sharply visible on a personal level in the short term. Humans are hard-wired for quick fight-or-flight reactions in the face of an imminent threat, but not highly motivated to act against slow-moving and somewhat abstract problems, even if the challenges that they pose are ultimately dire.
Data science involves people 🧍¶
The decisions that we make as data scientists have the potential to impact the livelihoods of other people.
- Flu case forecasting.
- Admissions and hiring.
- Hyper-personalized ad recommendations.
What is this course really about, then?¶
- Good data analysis is not:
- A simple application of a statistics formula.
- A simple application of computer programs.
- There are many tools out there for data science, but they are merely tools. They don’t do any of the important thinking – that's where you come in!
Course content¶
Course goals¶
DSC 80 teaches you to think like a data scientist.
In this course, you will...
- Get a taste of the "life of a data scientist."
- Practice translating potentially vague questions into quantitative questions about measurable observations.
- Learn to reason about "black-box" processes (e.g. complicated models).
- Understand computational and statistical implications of working with data.
- Learn to use real data tools (and rely on documentation).
Course outcomes¶
After this course, you will...
- Be prepared for internships and data science "take home" interviews!
- Be ready to create your own portfolio of personal projects.
- Have the background and maturity to succeed in the upper-division.
Topics¶
- Week 1: From
babypandas
topandas
. - Week 2: DataFrames.
- Week 3: Working with messy data, hypothesis and permutation testing.
- Week 4: Missing values.
- Week 5: HTML, Midterm Exam.
- Week 6: Web and text data.
- Week 7: Text data, modeling.
- Week 8: Feature engineering and generalization.
- Week 9: Modeling in
sklearn
. - Week 10: Classifier evaluation, fairness, conclusion.
- Week 11: Final Exam
Course logistics¶
Getting set up¶
- Ed: Q&A forum. Must be active here, since this is where all announcements will be made.
- Gradescope: Where you will submit all assignments for autograding, and where all of your grades will live.
- Canvas: No ❌.
In addition, you must fill out our Welcome Survey.
Accessing course content on GitHub¶
You will access all course content by pulling the course GitHub repository:
We will post HTML versions of lecture notebooks on the course website, but otherwise you must git pull
from this repository to access all course materials (including blank copies of assignments).
Environment setup¶
- You're required to set up a Python environment on your own computer.
- To do so, follow the instructions on the Tech Support page of the course website.
- Once you set up your environment, you will
git pull
the course repo every time a new assignment comes out. - Note: You will submit your work to Gradescope directly, without using Git.
- We will try to post a demo video with Lab 1, and we'll help you with this in Discussion 1 this week.
Lectures¶
- Lectures are held in-person on Tuesdays and Thursdays from 9:30am-11am and 11am-12:30pm in Peterson Hall 103.
Lectures are podcasted.
New for this quarter: Assignment deadlines are fairly flexible (as I'll explain soon). To help yourself stay on track with material, you can opt into lecture attendance. If you do, lecture attendance is worth 5% of your overall grade (instead of 0%) and the midterm and final are worth 2.5% less.
To get credit for a week, attend and participate in the activities for both lectures. Lowest two weeks dropped.
Assignments¶
In this course, you will learn by doing!
- Labs (20%): 9 total, lowest score dropped. Usually due weekly on Wednesdays at 11:59PM.
- Projects (25% + 5% checkpoints): 4 total, no drops. Usually due on Fridays at 11:59PM, and usually have a "checkpoint."
In DSC 80, assignments will usually consist of both a Jupyter Notebook and a .py
file. You will write your code in the .py
file; the Jupyter Notebook will contain problem descriptions and test cases. Lab 1 will explain the workflow.
Late Policy¶
- No late submissions accepted, but...
- Extension Request Form grants you 2 days extension on assignment submission (instead of slip days).
- We will essentially approve all requests!
- Goal is to help support you if you start falling behind, so if you fill out form a lot, we'll schedule a meeting with you to help come up with a plan for success.
- No extensions on project checkpoints or Final Project.
Redemption for Labs and Projects¶
- All labs and projects have hidden autograder tests.
- We won't show you what you missed until the deadline has passed.
- But you can resubmit after the original deadline to redeem up to 80% of points lost.
Discussions¶
- Most common feedback from past quarter: want more exam prep!
- Discussions will have a weekly worksheet with past exam questions.
- If you opt-in, discussion will be worth 5% of your overall grade (instead of 0%), and the midterm and final exams will be worth 2.5% less.
- Get credit by attending discussion and completing the worksheet. Lowest 2 weeks dropped.
Exams¶
- Midterm Exam (20%): Thursday, May 2nd, in-person during lecture.
- Final Exam (30%): Saturday, June 8th, 8AM-11AM in-person (location TBD).
- Your final exam score can redeem your midterm score (see the Syllabus for details).
- Let us know on the Exam Accommodations Request Form if you have a conflict.
A typical week in DSC 80¶
Resources¶
- Your main resource will be lecture notebooks.
- Most lectures also have supplemental readings that come from our course textbook, Learning Data Science. These are not required, but are highly recommended.
Support 🫂¶
It is no secret that this course requires a lot of work – becoming fluent with working with data is hard!
- You will learn how to solve problems independently – documentation and the internet will be your friends.
- Learning how to effectively check your work and debug is extremely useful.
- Learning to stick with a problem (tenacity) is a very valuable skill; but don't be afraid to ask for help.
Once you've tried to solve problems on your own, we're glad to help.
- We have several office hours in person each week. See the Calendar 📆 for details.
- Ed is your friend too. Make your conceptual questions public, and make your debugging questions private.
Generative Artificial Intelligence¶
- We know that tools, like ChatGPT and GitHub Copilot, can write code for you.
- Feel free to use such tools with caution. Refer to the Generative AI section of the syllabus for details.
- We trust that you're here to learn and do the work for yourself.
- You won't be able to use ChatGPT on the exams, which are in-person and on paper, so make sure you understand how your code actually works.
The data science lifecycle 🚴¶
The scientific method¶
You learned about the scientific method in elementary school.
However, it hides a lot of complexity.
- Where did the hypothesis come from?
- What data are you modeling? Is the data sufficient?
- Under which conditions are the conclusions valid?
The data science lifecycle¶
All steps lead to more questions! We'll refer back to the data science lifecycle repeatedly throughout the quarter.
Example: What's in a name?¶
Lilith, Lilibet … Lucifer? How Baby Names Went to 'L'¶
This New York Times article claims that baby names beginning with "L" have become more popular over time.
Let's see if these claims are true, based on the data!
The data¶
What we're seeing below is a pandas
DataFrame. The DataFrame contains one row for every combination of 'Name'
, 'Sex'
, and 'Year'
.
baby = pd.read_csv('data/baby.csv')
baby
Name | Sex | Count | Year | |
---|---|---|---|---|
0 | Liam | M | 20456 | 2022 |
1 | Noah | M | 18621 | 2022 |
2 | Olivia | F | 16573 | 2022 |
... | ... | ... | ... | ... |
2085155 | Wright | M | 5 | 1880 |
2085156 | York | M | 5 | 1880 |
2085157 | Zachariah | M | 5 | 1880 |
2085158 rows × 4 columns
Recall from DSC 10, to access columns in a DataFrame, you used the .get
method.
baby.get('Count').sum()
365296191
Everything you learned in babypandas
translates to pandas
. However, the more common way of accessing a column in pandas
involves dictionary syntax:
baby['Count'].sum()
365296191
You'll learn more about this in Thursday's lecture.
How many unique names were there per year?¶
baby.groupby('Year').count()
Name | Sex | Count | |
---|---|---|---|
Year | |||
1880 | 2000 | 2000 | 2000 |
1881 | 1934 | 1934 | 1934 |
1882 | 2127 | 2127 | 2127 |
... | ... | ... | ... |
2020 | 31517 | 31517 | 31517 |
2021 | 31685 | 31685 | 31685 |
2022 | 31915 | 31915 | 31915 |
143 rows × 3 columns
A shortcut to the above is as follows:
baby['Year'].value_counts()
2008 35094 2007 34966 2009 34724 ... 1883 2084 1880 2000 1881 1934 Name: Year, Length: 143, dtype: int64
Why doesn't the above Series actually contain the number of unique names per year?
baby[(baby['Year'] == 1880)]
Name | Sex | Count | Year | |
---|---|---|---|---|
2083158 | John | M | 9655 | 1880 |
2083159 | William | M | 9532 | 1880 |
2083160 | Mary | F | 7065 | 1880 |
... | ... | ... | ... | ... |
2085155 | Wright | M | 5 | 1880 |
2085156 | York | M | 5 | 1880 |
2085157 | Zachariah | M | 5 | 1880 |
2000 rows × 4 columns
baby[(baby['Year'] == 1880)].value_counts('Name')
Name Grace 2 Emma 2 Clair 2 .. Evaline 1 Evalena 1 Zula 1 Length: 1889, dtype: int64
How many babies were recorded per year?¶
baby.groupby('Year').sum()
Count | |
---|---|
Year | |
1880 | 201484 |
1881 | 192690 |
1882 | 221533 |
... | ... |
2020 | 3333981 |
2021 | 3379713 |
2022 | 3361896 |
143 rows × 1 columns
baby.groupby('Year').sum().plot()
"'L' has to be like the consonant of the decade."¶
(baby
.assign(first_letter=baby['Name'].str[0])
.query('first_letter == "L"')
.groupby('Year')
.sum()
.plot(title='Number of Babies Born with an "L" Name Per Year')
)
What about individual names?¶
(baby
.query('Name == "Siri"')
.groupby('Year')
.sum()
.plot(title='Number of Babies Born Named "Siri" Per Year')
)
def name_graph(name):
return (baby
.query(f'Name == "{name}"')
.groupby('Year')
.sum()
.plot(title=f'Number of Babies Born Named "{name}" Per Year')
)
name_graph('Sam')
What about other names?¶
Visit http://q.dsc80.com/ to respond.
name_graph(...)
This week...¶
- On Thursday, we'll do a deep dive into
pandas
. - Lab 1 will be released by tomorrow.
- Come to discussion Friday / Monday for help setting up your environment, which you'll need to do before working on Lab 1.
- Also fill out the Welcome Survey and read the Syllabus!