365 Data Science courses free until November 21

The initiative presents a risk-free way to break into data science and an opportunity to upskill for free

The online educational platform 365 Data Science launches its #21DaysFREE campaign, providing 100% free unlimited access to all its content for three weeks. From November 1 to 21, you can take courses from renowned instructors, complete exams, and earn industry-recognized certificates.



About the Platform

365 Data Science has helped over 2 million students worldwide gain skills and knowledge in data science, analytics, and business intelligence. The program offers an all-encompassing framework for studying data science—whether you’re starting from scratch or looking to upgrade your skills. Students learn with self-paced video lessons and a myriad of exercises and real-world examples. Moreover, 365’s new gamified platform makes the learning journey engaging and rewarding. Providing a theoretical foundation for all data-related disciplines- Probability, Statistics, and Mathematics, the program also offers a comprehensive introduction to R programming, statistics in R, and courses on data visualization in R.

 

#21DaysFREE Campaign

Until November 21, you can take as many courses as you want and add new tools to your analytics skill set for free. The platform unlocks all 195 hours of video lessons, hundreds of practical exercises, career tracks, exams, and the opportunity to earn industry-recognized certificates. “Starting a career in data science requires devotion and determination. Our mission is to give everyone a chance to get familiar with the field and help them succeed professionally,” says Ned Krastev, CEO of 365 Data Science. This isn’t 365’s first free initiative. The idea of providing unlimited access to all courses was born during the 2020 COVID-19 lockdowns. “We felt it was the right time to open our platform,” adds Ned. “We tried to help people who had lost their jobs or wanted to switch careers to make a transition into data science and analytics.” The free access initiative drove unprecedented levels of engagement, which inspired the 365 team to turn it into a yearly endeavor. Their 2021 campaign, in just one month, generated 80,000 new students (aspiring data scientists and analytics specialists) from 200 countries, who viewed 7.5 million minutes of educational content and earned 35,000 certificates.

While 21 days is not enough to become a fully-fledged professional, the #21DaysFREE initiative provides a risk-free way to familiarize yourself with the industry and lay the foundations of a successful career. Join the program and start for free at  365 Data Science  

 

Bayesian multilevel modeling in R with brms workshop

Learn how to use Bayesian multilevel modeling in R, while contributing to charity! Join our workshop on Bayesian multilevel modeling in R with brms that is a part of our workshops for Ukraine series. 

Here’s some more info: 
Title: Bayesian multilevel modeling in R with brms

Date: Thursday, November 10th, 18:00 – 20:00 CET (Rome, Berlin, Paris timezone)
Speaker:
Paul Bürkner is a statistician currently working as a Junior Research Group Leader at the Cluster of Excellence SimTech at the University of Stuttgart (Germany). He is interested in a wide range of research topics most of which involve the development, evaluation, implementation, or application of Bayesian methods.  He is the author of the R package brms and member of the Stan Development Team. Previously, Paul studied Psychology and Mathematics at the Universities of Münster and Hagen (Germany) and did his PhD in Münster about optimal design and Bayesian data analysis. He has also worked as a Postdoctoral researcher at the Department of Computer Science at Aalto University (Finland).
Description:
The workshop will be about Bayesian multilevel models and their implementation in R using the package brms. At start there will be a short introduction to multilevel modeling and to Bayesian statistics in general followed by an introduction to Stan, which is an incredibly flexible language to fit open-ended Bayesian models. I will then explain how to access Stan using just basic R formula syntax via the brms package. It supports a wide range of response distributions and modeling options such as splines, autocorrelation, or censoring all in a multilevel context. A lot of post-processing and plotting methods are implemented as well. Some examples from Psychology and Medicine will be discussed.
Minimal registration fee:
20 euro (or 20 USD or 750 UAH)

How can I register?
  • Go to https://bit.ly/3wvwMA6 or  https://bit.ly/3PFxtNA and donate at least 20 euro. Feel free to donate more if you can, all proceeds go directly to support Ukraine.
  • Save your donation receipt (after the donation is processed, there is an option to enter your email address on the website to which the donation receipt is sent)
  • Fill in the registration form, attaching a screenshot of a donation receipt (please attach the screenshot of the donation receipt that was emailed to you rather than the page you see after donation).

If you are not personally interested in attending, you can also contribute by sponsoring a participation of a student, who will then be able to participate for free. If you choose to sponsor a student, all proceeds will also go directly to organisations working in Ukraine. You can either sponsor a particular student or you can leave it up to us so that we can allocate the sponsored place to students who have signed up for the waiting list.

How can I sponsor a student?
  • Go to https://bit.ly/3wvwMA6 or https://bit.ly/3PFxtNA and donate at least 20 euro (or 17 GBP or 20 USD or 750 UAH). Feel free to donate more if you can, all proceeds go to support Ukraine!
  • Save your donation receipt (after the donation is processed, there is an option to enter your email address on the website to which the donation receipt is sent)
  • Fill in the sponsorship form, attaching the screenshot of the donation receipt (please attach the screenshot of the donation receipt that was emailed to you rather than the page you see after the donation). You can indicate whether you want to sponsor a particular student or we can allocate this spot ourselves to the students from the waiting list. You can also indicate whether you prefer us to prioritize students from developing countries when assigning place(s) that you sponsored.

If you are a university student and cannot afford the registration fee, you can also sign up for the waiting list here. (Note that you are not guaranteed to participate by signing up for the waiting list).

You can also find more information about this workshop series,  a schedule of our future workshops as well as a list of our past workshops which you can get the recordings & materials of here.
Looking forward to seeing you during the workshop!

How do I count thee? Let me count the ways?

How do I count thee? Let me count the ways? by Jerry Tuttle    In Major League Baseball, a player who hits 50 home runs in a single season has hit a lot of home runs. Suppose I want to count the number of 50 homer seasons by team, and also the number of 50 homer seasons by New York Yankees. (I will count Maris and Mantle in 1961 as two.) Here is the data including Aaron Judge’s 62 in 2022 :
You would think base R would have a count function such as count(df$Team) and count(df$Team == “NYY”) but this gives the error “could not find function ‘count'”. Base R does not have a count function. Base R has at last four ways to perform a count: 1. The table function will count items in a vector.    table(df$Team) presents results horizontally, and data.frame(table(df$Team)) presents results vertically.    table(df$Team == “NYY”) displays results 37 false and 10 true, while table(df$Team == “NYY”)[2] just displays the result 10 true. 2. The sum function can be used to count the number of rows meeting a condition.    sum(df$Team == “NYY”) displays the result 10. Here df$Team == “NYY” is creating a logical vector, and sum is summing the number of true = 1. 3. Similar to sum, nrow(df[df$Team == “NYY”, ]) counts the number of rows meeting the NYY condition. 4. The length function counts the number of elements in an R object.    length(which(df$Team == “NYY”)) , length(df$Team[df$Team == “NYY”]) , and length(grep(“NYY”, df[ , “Team”])) are all ways that will count the 10 Yankees. The more direct solution to counting uses the count function in the dplyr library. Note that dplyr’s count function applies to a data frame or tibble, but not to a vector. After loading library(dplyr) , 1. df %>% count(Team) lists the count for each team. 2. df %>% filter(Team = “NYY”) lists each Yankee, and you can see there are 10. 3. df %>% count(Team == “NYY”) displays 37 false and 10 true, while df %>% filter(Team == “NYY”) %>% count() just displays the 10 true. The following is a bar chart of the results by team for teams with at least 1 50 homer season:
Finally, “How do I count thee? Let me count the ways?” is of course adapted from Elizabeth Barrett Browning’s poem “How do I love thee? Let me count the ways?” But in her poem, just how would we count the number of times “love” is mentioned? The tidytext library makes counting words fairly easy, and the answer is ten, the same number of 50 homer Yankee seasons. Coincidence? The following is all the R code. Happy counting!

  library(dplyr) 
library(ggplot2)
library(tidytext)
df <- data.frame(
   Player=c('Ruth','Ruth','Ruth','Ruth','Wilson','Foxx','Greenberg','Foxx','Kiner','Mize','Kiner','Mays','Mantle','Maris', 'Mantle','Mays','Foster','Fielder','Belle','McGwire','Anderson','McGwire','Griffey','McGwire','Sosa','Griffey', 'Vaughn','McGwire','Sosa','Sosa','Bonds','Sosa','Gonzalez','Rodriguez','Rodriguez','Thome','Jones','Howard','Ortiz', 'Rodriguez','Fielder','Bautista','Davis','Stanton','Judge','Alonso','Judge'),
   Year=c(1920,1921,1927,1928,1930,1932,1938,1938,1947,1947,1949,1955,1956,1961,1961,1965,1977,1990,1995,1996,1996,1997,1997, 1998,1998,1998,1998,1999,1999,2000,2001,2001,2001,2001,2002,2002,2005,2006,2006,2007,2007,2010,2013,2017,2017,2019,2022),
   Homers=c(54,59,60,54,56,58,58,50,51,51,54,51,52,61,54,52,52,51,50,52,50,58,56,70,66,56,50,65,63,50,73,64,57,52,57,52,51, 58,54,54,50,54,53,59,52,53,62),
   Team=c('NYY','NYY','NYY','NYY','CHC','PHA','DET','BOS','PIT','NYG','PIT','NYG','NYY','NYY','NYY','SF','CIN','DET','CLE', 'OAK','BAL','OAK/SLC','SEA','SLC','CHC','SEA','SD','SLC','CHC','CHC','SF','CHC','ARI','TEX','TEX','CLE','ATL','PHP', 'BOR','NYY','MIL','TOR','BAL','MIA','NYY','NYM','NYY')) head(df) # base R ways to count: table(df$Team)    # shows results horizontally
data.frame(table(df$Team))    #shows results vertically
table(df$Team == "NYY")    # displays 37 false and 10 true
table(df$Team == "NYY")[2] sum(df$Team == "NYY")    # displays the result 10. nrow(df[df$Team == "NYY", ])    # counts the number of rows meeting the NYY condition. length(which(df$Team == "NYY"))     # which returns a vector of indices which are true
length(df$Team[df$Team == "NYY"])
length(grep("NYY", df[ , "Team"]))     # grep returns a vector of indices that match the pattern # dplyr R ways to count; remember to load library(dplyr): df %>% count(Team)    # lists the count for each team. df %>% filter(Team == "NYY")    # lists each Yankee, and you can see there are 10. df %>% count(Team == "NYY")    # displays 37 false and 10 true, while
df %>% filter(Team == "NYY") %>% count()    # just displays the 10 true. # barplot of all teams with at least 1 50 homer season; remember to load library(ggplot2) df %>%
    group_by(Team) %>%
    summarise(count = n()) %>%
    ggplot(aes(x=reorder(Team, count), y=count, fill=Team)) +
    geom_bar(stat = 'identity') +
    ggtitle("Count of 50 Homer Seasons") +
    xlab("Team") +
    scale_y_continuous(breaks=c(1,2,3,4,5,6,7,8,9,10)) +
    coord_flip() +
    theme(plot.title = element_text(face="bold", size=18)) +
    theme(axis.title.y = element_text(face="bold")) +
    theme(axis.title.x = element_blank()) +
    theme(axis.text.x = element_text(size=12, face="bold"),
    axis.text.y = element_text(size=12, face="bold")) +
    theme(legend.position="none")
# count number of times "love" is mentioned in Browning's poem; remember to load library(tidytext) textfile <- c("How do I love thee? Let me count the ways.",
"I love thee to the depth and breadth and height",
"My soul can reach, when feeling out of sight",
"For the ends of being and ideal grace.",
"I love thee to the level of every day's",
"Most quiet need, by sun and candle-light.",
"I love thee freely, as men strive for right.",
"I love thee purely, as they turn from praise.",
"I love thee with the passion put to use", <br
"In my old griefs, and with my childhood's faith.",
"I love thee with a love I seemed to lose",
"With my lost saints. I love thee with the breath,", <br
"Smiles, tears, of all my life; and, if God choose,",
"I shall but love thee better after death.")
df<-data.frame(line=1:length(textfile), text=textfile) <br <br
df_words % unnest_tokens(word, text) <br
cleaned_words % anti_join(get_stopwords())
cleaned_words %>% count(word, sort = TRUE) %>% head(6)
cleaned_words %>% filter(word == "love") %>% count()

Unlock Your Data-Driven Future with DataCamp’s Analyst Takeover

For the rest of October, DataCamp is spotlighting how analyst skills transform careers and organizations. Featuring top courses, tracks and certifications for all levels, expect podcasts, webinars, and more from industry leaders—breaking down how the latest in R, Power BI, Tableau, Spreadsheets, and SQL.

DataCamp will be focussing on how being data-driven requires not only investments in tools and technology, but in people’s skills and an organization’s data culture. Illustrating how at a truly data-driven business, everyone adopts a common language, feels empowered to ask the right questions, and are able to make decisions based on facts—not just instinct.
The time to invest in analyst skills is now Only 26% of data leaders claim their organization is “data-driven”, whilst 88% or organizations those who have invested in organization-wide data upskilling programs have exceeded business goals. (Deloitte, 2022)
So, how do we get there? Look no further than DataCamp and their hands-on and interactive R tracks for supercharging your data journey.
Additionally, the majority of the introduction courses are completely free—allowing you to get a flavor of the technology and track before committing to a subscription.

CAREER TRACK | Data Analyst with R—no experience required.


Grow your R skills, and begin your journey to becoming a confident data analyst! No prior coding experience required. In this track, you’ll learn how to import, clean, manipulate, and visualize data in R—all integral skills for any aspiring data professional or researcher.
Across 9 courses taking approximately 36 hours you’ll run through interactive exercises, get hands-on with some of the most popular R packages, including ggplot2 and tidyverse packages like dplyr and readr. Develop your data manipulation and exploratory data analysis by working with real-world datasets, including everything from U.S. income data to global food consumption. 
Start Track Now
If you’re just looking for a short taster into what R is and if it could potentially be useful to you or your business, DataCamp’s free Introduction to R course (the first course on the Data Analyst with R Track) is just four hours. During the beginner-friendly courses, you will cover the basics of the open source language, including vectors, factors, lists, and data frames. In an extremely short time, you’ll gain useful coding skills and be ready to start your own data analysis in R.
Looking for something more advanced? 
If you’re ready to take your next career step—look no further. Check out the options below, or head straight to DataCamp’s R Tack Hub to scroll through a wider list

CAREER TRACK | Machine Learning Scientist with R

The essential skills to land a job as a machine learning scientist.
Here, you’ll augment your R programming skills with the toolbox to perform supervised and unsupervised learning. Learn how to process data for modeling, train your models, visualize your models and assess their performance, and tune their parameters for better performance. In the process, you’ll get an introduction to Bayesian statistics, natural language processing, and Spark.
Transform your career trajectory in just 57 hours, across 14 courses, of interactive learning… What are you waiting for?
Start Track Now

CAREER TRACK | Quantitative Analyst with R
Those who are ready to take on advanced topics. Over 15 courses, learn the ins and outs of quantitative analysis with R—invaluable skills for the finance sector. In finance, quantitative analysts ensure portfolios are risk balanced, help find new trading opportunities, and evaluate asset prices using mathematical models.
Go from learning essential data structures such as lists and data frames and apply that knowledge directly to financial examples in Introduction to R for Finance and Intermediate R for Finance to Bond Valuation and Analysis in R and beyond.
Change your career in just 65 hours.
Start Track Now
Certifications:
When you’re ready for the next step, DataCamp’s Certification Program validates new data skills to employers and recruiters. Listed as #1 in Forbes’s list of best data analytics certifications, DataCamp’s certifications are most respected within the data industry—helping learners fast-track hiring processes and helping new data professionals stand out.
Discover More Here


About DataCamp Whatever your starting point or data goals, DataCamp is your one-stop-shop for all your data needs. Across 380+ courses, track and certifications for all in levels—DataCamp’s hands-on learning takes complete beginners to certified data professionals, then helps their learners find a job to match their new skills.
Right now, DataCamp for Business is offering up to 60% off their entire platform for a limited time only. With a 23% increase in the number of job openings predicted for analysts between now and 2031 there has never been a better time to upskill in data skills (US Bureau of Labor Statistics, 2022).
Join over 10 million learners, 2,500+ companies and 80% of the Fortune 1000 who use DataCamp to upskill their teams—you’re in good hands.
Discover More.

Classification modeling in R for profitable decisions workshop

Learn classification modeling to improve your decision-making for your business or use these skills for your research in our 2-part workshop! These workshops are a part of our workshops for Ukraine series, and all proceeds from these workshops go to support Ukraine. You can find more information about other workshops, as well as purchase recordings of the previous workshops here.

In the first part of the workshop titled Classification modeling for profitable decisions, which will take place online on Thursday, October 20th, 18:00 – 20:00 CET, we will cover the theoretical framework that you need to know to perform classification analysis and cover the key concepts. 

The second part of the workshop that will take place on Thursday, October 27th, 18:00 – 20:00 CET will include hands-on practice in R, so that you can learn how to implement the concepts covered in the first part in R. 

You can register for each part separately, so you can choose whether you wish to attend both parts or just part 1 or part 2.   Below you can find more information about each part and how to register for it: 

PART 1
Title: Classification modeling for profitable decisions: Theory and a case study on firm defaults. 
Date:
Thursday, October 20th, 18:00 – 20:00 CET (Rome, Berlin, Paris timezone)
Speaker:
Gábor Békés is an Assistant Professor at the Department of Economics and Business of Central European University, a research fellow at KRTK in Hungary, and a research affiliate at CEPR. His research is focused on international economics; economic geography and applied IO, and was published among others by the Global Strategy Journal, Journal of International Economics, Regional Science and Urban Economics or Economic Policy and have authored commentary on VOXEU.org. His comprehensive textbook, Data Analysis for Business, Economics, and Policy with Gábor Kézdi was publsihed by Cambridge University Press in 2021. 
Description:
This workshop will introduce the framework and methods of probability prediction and classification analysis for binary target variable. We will discuss the key concepts such as probability prediction, classification threshold, loss function, classification, confusion table, expected loss, the ROC curve, AUC and more. We will use logit models as well as random forest to predict probabilities and classify. In the workshop we will focus on a case study on firm defaults using a dataset on financial and management features of firms. The workshop material is based on a chapter and a case study from my textbook. Code in R and Python are available from the Github repo, and the data is available as well. The workshop will introduce key concepts, but the focus will be on data wrangling and modelling decisions we make for a real life problem. There will be a follow-up workshop focusing on the coding side of the case study. 
Minimal registration fee:
20 euro (or 20 USD or 750 UAH)
Suggested registration fee for professionals:
50 euro (if you can afford it, our suggested registration fee for this workshop is 50 euro. If you cannot afford it, you can still register by donating 20 euro).

Remember that you can register even if you will not be able to attend in person as all registered participants will get a recording.

How can I register?
  • Go to https://bit.ly/3wvwMA6 or https://bit.ly/3PFxtNA and donate at least 20 euro. Feel free to donate more if you can, all proceeds go to support Ukraine!
  • Save your donation receipt (after the donation is processed, there is an option to enter your email address on the website to which the donation receipt is sent)
  • Fill in the registration form, attaching a screenshot of a donation receipt (please attach the screenshot of the donation receipt that was emailed to you rather than the page you see after donation).

If you are not personally interested in attending, you can also contribute by sponsoring a participation of a student, who will then be able to participate for free. If you choose to sponsor a student, all proceeds will also go directly to organisations working in Ukraine. You can either sponsor a particular student or you can leave it up to us so that we can allocate the sponsored place to students who have signed up for the waiting list.

How can I sponsor a student?
  • Go to https://bit.ly/3wvwMA6 or https://bit.ly/3PFxtNA and donate at least 20 euro (or 17 GBP or 23 USD or 660 UAH). Feel free to donate more if you can, all proceeds go to support Ukraine!
  • Save your donation receipt (after the donation is processed, there is an option to enter your email address on the website to which the donation receipt is sent)
  • Fill in the sponsorship form, attaching the screenshot of the donation receipt (please attach the screenshot of the donation receipt that was emailed to you rather than the page you see after the donation). You can indicate whether you want to sponsor a particular student or we can allocate this spot ourselves to the students from the waiting list. You can also indicate whether you prefer us to prioritize students from developing countries when assigning place(s) that you sponsored.

If you are a university student and cannot afford the registration fee, you can also sign up for the waiting list here. (Note that you are not guaranteed to participate by signing up for the waiting list).


PART 2
Title: Classification modelling for profitable decisions: Hands on practice in R
Date:
Thursday, October 27th, 18:00 – 20:00 CET (Rome, Berlin, Paris timezone)
Speaker:
Ágoston Reguly is a Postdoctoral Fellow at the Financial Services and Innovation Lab of Scheller College of Business, Georgia Institute of Technology. His research is focused on causal machine learning methods and their application in corporate finance. He obtained his Ph.D. degree from Central European University (CEU), where he has taught multiple courses such as data analysis, coding, and mathematics. Before CEU he worked for more than three years at the Hungarian Government Debt Management Agency.
Description:
This workshop will implement methods of probability prediction and classification analysis for the binary target variable. This workshop is a follow-up to Gábor Békés’s workshop on the key concepts and (theoretical) methods for the same subject. We will use R via RStudio to apply probability prediction, classification threshold, loss function, classification, confusion table, expected loss, the ROC curve, AUC, and more. We will use linear probability models, logit models as well as random forests to predict probabilities and classify. In the workshop, we follow the case study on firm defaults using a dataset on financial and management features of firms. The workshop material is based on a chapter and a case study from the textbook of Gábor Békés and Gábor Kézdi (2021): Data Analysis for Business, Economics, and Policy, Cambridge University Press. The workshop will not only implement the key concepts, but the focus will be on data wrangling and modeling decisions we make for a real-life problem. Minimal registration fee: 20 euro (or 20 USD or 750 UAH) Suggested registration fee for professionals: 50 euro (if you can afford it, our suggested registration fee for this workshop is 50 euro. If you cannot afford it, you can still register by donating 20 euro).

How can I register?
  • Go to https://bit.ly/3wvwMA6 or https://bit.ly/3PFxtNA and donate at least 20 euro. Feel free to donate more if you can, all proceeds go to support Ukraine!
  • Save your donation receipt (after the donation is processed, there is an option to enter your email address on the website to which the donation receipt is sent)
  • Fill in the registration form, attaching a screenshot of a donation receipt (please attach the screenshot of the donation receipt that was emailed to you rather than the page you see after donation).

How can I sponsor a student?
  • Go to https://bit.ly/3wvwMA6 or https://bit.ly/3PFxtNA and donate at least 20 euro (or 17 GBP or 23 USD or 660 UAH). Feel free to donate more if you can, all proceeds go to support Ukraine!
  • Save your donation receipt (after the donation is processed, there is an option to enter your email address on the website to which the donation receipt is sent)
  • Fill in the sponsorship form, attaching the screenshot of the donation receipt (please attach the screenshot of the donation receipt that was emailed to you rather than the page you see after the donation). You can indicate whether you want to sponsor a particular student or we can allocate this spot ourselves to the students from the waiting list. You can also indicate whether you prefer us to prioritize students from developing countries when assigning place(s) that you sponsored.

If you are a university student and cannot afford the registration fee, you can also sign up for the waiting list here. (Note that you are not guaranteed to participate by signing up for the waiting list).

Looking forward to seeing you during the workshop!

Access Datacamp’s top programming courses for only $1 this week!



Running only until Oct. 7, DataCamps’ entire learning platform is accessible for just $1 during Space Week!
Suitable for complete beginners and seasoned practitioners alike, DataCamp’s hands-on learning approach has something for anyone looking to advance their data skills. With unlimited access to 380+ courses in Python, R, SQL, Power-bi and more for only $1—you can start coding and applying your skills in real-world situations from day one, regardless of your starting point. And, once you’ve mastered certain skills, we offer industry-leading certifications to help boost your career and get recognised by recruiters. As data becomes more prominent in our daily and working lives, possessing the skills to effectively use and manipulate data puts you at an instant advantage. In fact, 82% of leaders expect all employees to have basic data literacy skills (Tableau, 2022)—and it’s only going to increase. For less than a price of coffee, dive into data. Your future self will thank you. Buy Now *note: First month is charged at $1, after which you will be billed monthly at our standard rate. Cancel any time.

TidyFinance: Empirical asset pricing in R workshop

Learn how to do empirical asset pricing in R, while contributing to charity! Join our workshop on TidyFinance: Empirical asset pricing in R which is a part of our workshops for Ukraine series. 

Here’s some more info: 
Title: TidyFinance: Empirical asset pricing in R
Date:
Thursday, October 13th 18:00 – 20:00 CEST (Rome, Berlin, Paris timezone)
Speaker:
Patrick Weiss, PhD, CFA is a postdoctoral researcher at Vienna University of Economics and Business. Jointly with Christoph Scheuch and Stefan Voigt, Patrick wrote the open-source book www.tidy-finance.org , which serves as the basis for this workshops. Visit his webpage for additional information.
Description:
This workshop explores empirical asset pricing and combines explanations of theoretical concepts with practical implementations. The course relies on material available on www.tidy-finance.org and proceeds in three steps: (1) We dive into the most used data sources and show how to work with data from WRDS, forming the basis for the analysis. We also briefly introduce some other possible sources of financial data. (2) We show how to implement the capital asset pricing model in rolling-window regressions. (3) We introduce the widely used method of portfolio sorts in empirical asset pricing. During the workshop, we will combine some theoretical insights with hands-on implementations in R.


How can I register?

  • Go to https://bit.ly/3PFxtNA and donate at least 20 euro. Feel free to donate more if you can, all proceeds go directly to support Ukraine.
  • Save your donation receipt (after the donation is processed, there is an option to enter your email address on the website to which the donation receipt is sent)
  • Fill in the registration form, attaching a screenshot of a donation receipt (please attach the screenshot of the donation receipt that was emailed to you rather than the page you see after donation).

If you are not personally interested in attending, you can also contribute by sponsoring a participation of a student, who will then be able to participate for free. If you choose to sponsor a student, all proceeds will also go directly to organisations working in Ukraine. You can either sponsor a particular student or you can leave it up to us so that we can allocate the sponsored place to students who have signed up for the waiting list.

How can I sponsor a student?
  • Go to https://bit.ly/3PFxtNA and donate at least 20 euro (or 17 GBP or 20 USD or 750 UAH). Feel free to donate more if you can, all proceeds go to support Ukraine!
  • Save your donation receipt (after the donation is processed, there is an option to enter your email address on the website to which the donation receipt is sent)
  • Fill in the sponsorship form, attaching the screenshot of the donation receipt (please attach the screenshot of the donation receipt that was emailed to you rather than the page you see after the donation). You can indicate whether you want to sponsor a particular student or we can allocate this spot ourselves to the students from the waiting list. You can also indicate whether you prefer us to prioritize students from developing countries when assigning place(s) that you sponsored.

If you are a university student and cannot afford the registration fee, you can also sign up for the waiting list here. (Note that you are not guaranteed to participate by signing up for the waiting list).

You can also find more information about this workshop series,  a schedule of our future workshops as well as a list of our past workshops which you can get the recordings & materials of here.
Looking forward to seeing you during the workshop!

DataCamp Recruit: A better way to hire data professionals

DataCamp Recruit is built to help you find, hire and scale industry leading data teams. The platform provides access to one of the largest sources of certified data professionals, with clear insights into the precise skills, experience, and expertise that you need to hire for.

Why we launched DataCamp Recruit 

The demand for data professionals has never been higher, while the supply of qualified candidates remains low. On top of that, the difficulty around testing for technical skills makes hiring data professionals one of the most arduous roles for a recruiter.

With DataCamp Recruit, we provide recruiters with access to not only one of the largest sources of job-ready data professionals, but also clear insights into the technical abilities of each candidate. With our easy to use filters, you can get to the right hire faster.

What can DataCamp Recruit do for you? 

It can take months to find and hire the right data professional, but with DataCamp Recruit we get you there in minutes. 

Data professionals on DataCamp have a diverse set of experience and backgrounds. Through our filtering tool, you can match with candidates and skill sets that you need to hire for. This saves you time, and gives you clear insights into their skills, helping you to hire with confidence.  

DataCamp’s content and curriculum are designed by leaders in data science, so candidates are equipped with up-to-date technical skills making them ready to work from day one.

Plus you can hire candidates certified by our #1 ranked certification program (Forbes, 2022). Candidates certified by DataCamp are not only tested for their technical skills, but also on their soft skills such as their ability to translate data into insight, while communicating effectively.

Professionals on DataCamp Recruit are ranked across six key categories:

    • Data management
    • Exploratory analysis
    • Statistical experimentation
    • Model development
    • Coding in production environments
    • Communication

What candidate profiles on DataCamp Recruit look like:


Check out our short demo to get an idea of how to use DataCamp Recruit

[youtube https://www.youtube.com/watch?v=XWeMz4M0a8Q&w=560&h=315]

Get Started with Recruit

Like what you see? Get started with Recruit today for free, or request a free demo to find out more about the platform.



Fundamentals of Exploratory and Inferential Spatial Data Analysis in R workshop

Learn how to work with Spatial Data in R, while contributing to charity! Join our workshop on Fundamentals of Exploratory and Inferential Spatial Data Analysis in R which is a part of our workshops for Ukraine series. 

Here’s some more info: 

Title: Fundamentals of Exploratory and Inferential Spatial Data Analysis in R
Date: 
Thursday, September 13th, 18:00 – 20:00 CEST (Rome, Berlin, Paris timezone)
Speaker:
Denys Dukhovnov, Ph.D. student in Demography at University of California, Berkeley. His research revolves around small-area estimation and geographic inequalities in mortality in the United States. He holds a previous M.A. degree in Data Analytics and Applied Social Research, held multiple research positions in social science fields, and currently works as a researcher at the Human Mortality Database (HMD).
Description:
This workshop will provide a hands-on overview of the exploratory and inferential spatial data analysis in R. The attendees will become familiar with statistical concepts of spatial adjacency and dependence and with various methods of measuring it (using such indicators as Moran’s I, Geary’s C, LISA/ELSA plots, etc.), as well as with statistical challenges of working with spatial data (e.g. modifiable areal unit problem or MAUP). In addition, the workshop will provide a foundational overview of inferential spatial analysis, specifically through the application of the basic types of spatial econometric regression models (SAR, SLX, SEM models). An emphasis will be made on the interpretation and reporting of the model performance and results. Prior familiarity with spatial data types and OLS regression is helpful, but not necessary.

How can I register?
  • Go to https://bit.ly/3wvwMA6 or  https://bit.ly/3PFxtNA and donate at least 20 euro. Feel free to donate more if you can, all proceeds go directly to support Ukraine.
  • Save your donation receipt (after the donation is processed, there is an option to enter your email address on the website to which the donation receipt is sent)
  • Fill in the registration form, attaching a screenshot of a donation receipt (please attach the screenshot of the donation receipt that was emailed to you rather than the page you see after donation).

If you are not personally interested in attending, you can also contribute by sponsoring the participation of a student, who will then be able to participate for free. If you choose to sponsor a student, all proceeds will also go directly to organizations working in Ukraine. You can either sponsor a particular student or you can leave it up to us so that we can allocate the sponsored place to students who have signed up for the waiting list.

How can I sponsor a student?
  • Go to https://bit.ly/3wvwMA6 or https://bit.ly/3PFxtNA and donate at least 20 euro (or 17 GBP or 20 USD or 750 UAH). Feel free to donate more if you can, all proceeds go to support Ukraine!
  • Save your donation receipt (after the donation is processed, there is an option to enter your email address on the website to which the donation receipt is sent)
  • Fill in the sponsorship form, and attaching the screenshot of the donation receipt (please attach the screenshot of the donation receipt that was emailed to you rather than the page you see after the donation). You can indicate whether you want to sponsor a particular student or we can allocate this spot ourselves to the students from the waiting list. You can also indicate whether you prefer us to prioritize students from developing countries when assigning place(s) that you sponsored.

You can also find more information about this workshop series,  a schedule of our future workshops as well as a list of our past workshops which you can get the recordings & materials of here.

Looking forward to seeing you during the workshop!

How to generate data from a model – Part 1


Summary

Traditionally, data scientists have built models based on data. This article details how to do the exact opposite i.e. generate data based on a model. This article is first in the series of articles on building data from model. 


Motivation & Practical Applications

Businesses across various industry domains are embracing Artificial Intelligence(AI) and Machine Learning(ML) driven solutions. Furthermore, it is observed that the recent increase in the cloud based Machine Learning Operations (ML Ops) tools such Azure ML has made AI/ML solutions relatively more accessible,  easy to deploy and in some cases more affordable. Additionally, it is also observed that there is an increase in the usage of Auto ML & feature engineering packages. These approaches reduce manual intervention during model build and retraining stages.  Since the focus is predominantly on building ML pipelines as opposed to the traditional approach of building models manually, the robustness of the pipelines needs to be inspected. This is still an evolving field and currently is being handled by model observability tools. This article proposes one such method of observability. The purpose of this method can be best represented in the form of a question as given below.

What if we built the underlying data distributions, the outliers, the dependent variable and then put it through the ML Ops pipeline?  Wouldn’t we know where the pipeline worked well and where it could have done better?
This question motivated the build of a Software as a Service (SaaS) product called uncovr. This product can now be accessed through an R package conjurer by following the steps outlined below.


Data from Model Using R

Step 1: Register for the API

    • Head over to the API developer portal at (https://foyi.developer.azure-api.net/).
    • Click on the Sign up button on the home page.
    • Register your details such as email, password etc.
    • You will receive an email to verify your email id.
    • Once you verify your email id, your account will be setup and you will receive a confirmation email.
    • Once your account is set up, please head over to the products section on the developer portal and select the product starter. Currently, this is the only subscription available. Give your subscription a name, read and accept the terms and click Subscribe
    • On your profile page, under the subscriptions section, click on show next to the Primary key. That is the subscription key you will need to access the API.

Step 2: Install R Package Conjurer

Install the latest version of the package from CRAN as follows.
install.packages("conjurer")

Step 3: Generate data

Generate the data using the code below.

library(conjurer)
uncovrJson <- buildModelData(numOfObs = 1000, numOfVars = 3, key = "input your subscription key here")
df <- extractDf(uncovrJson=uncovrJson)

The above code has two steps. The first step is to connect to the API and source the data in JSON  format. The second step is to convert the JSON format to an R dataframe.

The components of the function buildModelData are as follows.
    • numOfObs is the number of observations i.e. rows of data that you would like to generate. Please note that the current version allows you to generate from a minimum of 100 observations to a maximum of 10,000. 
    • numOfVars is the number of independent variables i.e. columns in the data. Please note that the current version allows you to generate from a minimum of 1 variable to a maximum of 100.
    • key is the Primary key that you have sourced from the earlier step.
The data frame df (in the code above) will have three columns with the names iv1, iv2, iv3 and one column dv. The columns with prefix iv are the independent variables while the dv is the dependent variable. You can rename them to suit your needs. 
The model used in the current version to generate the data is a linear regression model. The details of the model formula and its estimated performance can be inspected as follows. 
    • To begin with, you can inspect the JSON data that is received from the API by using the code  str(uncovrJson). This should display all the components of the JSON file. The attributes prefixed as slope are the coefficients of the model formula corresponding to the number. For example, slope1 is the coefficient corresponding to iv1 i.e. independent variable 1. 
    • The regression formula used to construct the data for the example data frame is as follows.
      Please note that the formula takes the form of Y = mX + C.
      dv = intercept + (slope1*iv1) + (slope2*iv2) + (slope3*iv3) + error.
    • Please note that while the slopes i.e. the coefficients are at variable level, the error is at each observation level. These errors can be accessed as uncovrJson$error

Concluding Remarks

The underlying API uncovr is under development and as new functionality is released, the R package conjurer will be updated to reflect those changes. For any feature requests or bug reports, please follow the contribution guidelines on GitHub repository. If you would like to follow the future releases and news, please follow our LinkedIn page.