Learn machine learning the fun way, with Oracle and RedBull Racing

Overview

Red Bull Racing Analytics Hands-On Labs

License: UPL Quality gate

Introduction

Are you interested in learning machine learning (ML)? How about doing this in the context of the exciting world of F1 racing?! Get your ML skills bootstrapped here with Oracle and Red Bull Racing!

Red Bull F1 Race Car

This tutorial teaches ML analytics with a series of hands-on labs (HOLs) using the Data Science service in Oracle Cloud Infrastructure.

You'll learn how to get data from some public data sources, then how to analyze this data using some of the latest ML techniques. In the process you'll build ML models and test them out in a predictor app.

Getting Started

There is some infrastructure that must be deployed before you can enjoy this tutorial. See the Terraform documentation for more information.

After the OCI infrastructure is deployed, proceed with the beginner's tutorial to start through the ML labs.

Prerequisites

You must have an OCI account. Click here to create a new cloud account.

This solution is designed to work with several OCI services, allowing you to quickly be up-and-running:

There are required OCI resources (see the Terraform documentation for more information) that are needed for this tutorial.

Notes/Issues

None at this time.

URLs

Contributing

This project is open source. Please submit your contributions by forking this repository and submitting a pull request! Oracle appreciates any contributions that are made by the open source community.

License

Copyright (c) 2021 Oracle and/or its affiliates.

Licensed under the Universal Permissive License (UPL), Version 1.0.

See LICENSE for more details.

Comments
  • Refactored Terraform code

    Refactored Terraform code

    • Compatible with ORM, Cloud Shell and Terraform CLI
    • Updated README to include instructions for all three methods
    • Refactored, removing unnecessary resources (Vault, public Subnet, etc.).
    • Added a nerd knob so that it could use an existing Group (rather than create a new one)
    • Fixed ORM RegEx filters to allow dashes (-) and underscores (_), for the names
    opened by timclegg 2
  • Issue with hands on lab guide - launchapp.sh missing

    Issue with hands on lab guide - launchapp.sh missing

    https://github.com/oracle-devrel/redbull-analytics-hol/tree/main/beginners#beginners-hands-on-lab

    In Starting The Web Application it reads:

    cd /home/opc/redbull-analytics-hol/beginners/web ./launchapp.sh start

    However is launchapp.sh is missing, for example

    (redbullenv) cd /home/opc/redbull-analytics-hol/beginners/web (redbullenv) ./launchapp.sh start bash: ./launchapp.sh: No such file or directory

    opened by raekins 1
  • fix: Updating schema.yaml syntax

    fix: Updating schema.yaml syntax

    Making the variable notation follow what the doc syntax shows (https://docs.oracle.com/en-us/iaas/Content/ResourceManager/Concepts/terraformconfigresourcemanager_topic-schema.htm)

    opened by timclegg 1
  • Exploratory Data Analysis Merge Issue

    Exploratory Data Analysis Merge Issue

    Hello I have been encountering an issue while running the lab. The Jupyter notebook 03.f1_analysis_EDA.ipynb has the following issue on cell number 5:


    ValueError Traceback (most recent call last) in ----> 1 df1 = pd.merge(races,results,how='inner',on=['raceId']) 2 df2 = pd.merge(df1,quali,how='inner',on=['raceId','driverId','constructorId']) 3 df3 = pd.merge(df2,drivers,how='inner',on=['driverId']) 4 df4 = pd.merge(df3,constructors,how='inner',on=['constructorId']) 5 df5 = pd.merge(df4,circuit,how='inner',on=['circuitId'])

    ~/redbullenv/lib64/python3.6/site-packages/pandas/core/reshape/merge.py in merge(left, right, how, on, left_on, right_on, left_index, right_index, sort, suffixes, copy, indicator, validate) 85 copy=copy, 86 indicator=indicator, ---> 87 validate=validate, 88 ) 89 return op.get_result()

    ~/redbullenv/lib64/python3.6/site-packages/pandas/core/reshape/merge.py in init(self, left, right, how, on, left_on, right_on, axis, left_index, right_index, sort, suffixes, copy, indicator, validate) 654 # validate the merge keys dtypes. We may need to coerce 655 # to avoid incompatible dtypes --> 656 self._maybe_coerce_merge_keys() 657 658 # If argument passed to validate,

    ~/redbullenv/lib64/python3.6/site-packages/pandas/core/reshape/merge.py in _maybe_coerce_merge_keys(self) 1163 inferred_right in string_types and inferred_left not in string_types 1164 ): -> 1165 raise ValueError(msg) 1166 1167 # datetimelikes must match exactly

    ValueError: You are trying to merge on object and int64 columns. If you wish to proceed you should use pd.concat

    I’m using an oracle automatic deployment provided by oracle as part of their environment. I do not have a lot of experience with Python but one possible ible solution is to read the numeric values form the csv file as integer or float but I’m almost certain the solution might be a little more elaborated than that 😉. Anyway thanks for your time. I’m really excited to test your solution and finish the lab. Thanks again.

    opened by yankodavila 2
  • Has the PAR for the stack deploy image expired.

    Has the PAR for the stack deploy image expired.

    Cannot deploy stack as getting PAR expired message.

    2021/11/07 10:50:11[TERRAFORM_CONSOLE] [INFO] Error Message: work request did not succeed, workId: ocid1.coreservicesworkrequest.oc1.eu-amsterdam-1.abqw2ljrwz2n7qqj7ghdwtnlrqol355oumc7a6coushvgdrebskspaewh7ea, entity: image, action: CREATED. Message: Import image not found: PAR is invalid (maybe is expired or deleted), please check.

    PAR in stack file is https://objectstorage.eu-frankfurt-1.oraclecloud.com/p/khhPjc_IMuyBOMfZUcJajIzCpoZ5aC-D7VMCU__GVZRlIQueXLIIcaaqLOZIuT1a/n/emeasespainsandbox/b/publichol/o/redbullhol-20210809-1523

    opened by Mel-A-M 1
Releases(v0.1.8)
Owner
Oracle DevRel
Oracle DevRel
A Big Data ETL project in PySpark on the historical NYC Taxi Rides data

Processing NYC Taxi Data using PySpark ETL pipeline Description This is an project to extract, transform, and load large amount of data from NYC Taxi

Unnikrishnan 2 Dec 12, 2021
This repo contains a simple but effective tool made using python which can be used for quality control in statistical approach.

This repo contains a powerful tool made using python which is used to visualize, analyse and finally assess the quality of the product depending upon the given observations

SasiVatsal 8 Oct 18, 2022
Toolchest provides APIs for scientific and bioinformatic data analysis.

Toolchest Python Client Toolchest provides APIs for scientific and bioinformatic data analysis. It allows you to abstract away the costliness of runni

Toolchest 11 Jun 30, 2022
An ETL Pipeline of a large data set from a fictitious music streaming service named Sparkify.

An ETL Pipeline of a large data set from a fictitious music streaming service named Sparkify. The ETL process flows from AWS's S3 into staging tables in AWS Redshift.

1 Feb 11, 2022
Business Intelligence (BI) in Python, OLAP

Open Mining Business Intelligence (BI) Application Server written in Python Requirements Python 2.7 (Backend) Lua 5.2 or LuaJIT 5.1 (OML backend) Mong

Open Mining 1.2k Dec 27, 2022
Mortgage-loan-prediction - Show how to perform advanced Analytics and Machine Learning in Python using a full complement of PyData utilities

Mortgage-loan-prediction - Show how to perform advanced Analytics and Machine Learning in Python using a full complement of PyData utilities. This is aimed at those looking to get into the field of D

Joachim 1 Dec 26, 2021
Streamz helps you build pipelines to manage continuous streams of data

Streamz helps you build pipelines to manage continuous streams of data. It is simple to use in simple cases, but also supports complex pipelines that involve branching, joining, flow control, feedbac

Python Streamz 1.1k Dec 28, 2022
Python Library for learning (Structure and Parameter) and inference (Statistical and Causal) in Bayesian Networks.

pgmpy pgmpy is a python library for working with Probabilistic Graphical Models. Documentation and list of algorithms supported is at our official sit

pgmpy 2.2k Dec 25, 2022
Active Learning demo using two small datasets

ActiveLearningDemo How to run step one put the dataset folder and use command below to split the dataset to the required structure run utils.py For ea

3 Nov 10, 2021
In this tutorial, raster models of soil depth and soil water holding capacity for the United States will be sampled at random geographic coordinates within the state of Colorado.

Raster_Sampling_Demo (Resulting graph of this demo) Background Sampling values of a raster at specific geographic coordinates can be done with a numbe

2 Dec 13, 2022
Using Data Science with Machine Learning techniques (ETL pipeline and ML pipeline) to classify received messages after disasters.

Using Data Science with Machine Learning techniques (ETL pipeline and ML pipeline) to classify received messages after disasters.

1 Feb 11, 2022
Stock Analysis dashboard Using Streamlit and Python

StDashApp Stock Analysis Dashboard Using Streamlit and Python If you found the content useful and want to support my work, you can buy me a coffee! Th

StreamAlpha 27 Dec 09, 2022
ICLR 2022 Paper submission trend analysis

Visualize ICLR 2022 OpenReview Data

Jintang Li 75 Dec 06, 2022
PyChemia, Python Framework for Materials Discovery and Design

PyChemia, Python Framework for Materials Discovery and Design PyChemia is an open-source Python Library for materials structural search. The purpose o

Materials Discovery Group 61 Oct 02, 2022
A real-time financial data streaming pipeline and visualization platform using Apache Kafka, Cassandra, and Bokeh.

Realtime Financial Market Data Visualization and Analysis Introduction This repo shows my project about real-time stock data pipeline. All the code is

6 Sep 07, 2022
Spaghetti: an open-source Python library for the analysis of network-based spatial data

pysal/spaghetti SPAtial GrapHs: nETworks, Topology, & Inference Spaghetti is an open-source Python library for the analysis of network-based spatial d

Python Spatial Analysis Library 203 Jan 03, 2023
A CLI tool to reduce the friction between data scientists by reducing git conflicts removing notebook metadata and gracefully resolving git conflicts.

databooks is a package for reducing the friction data scientists while using Jupyter notebooks, by reducing the number of git conflicts between different notebooks and assisting in the resolution of

dataroots 86 Dec 25, 2022
This tool parses log data and allows to define analysis pipelines for anomaly detection.

logdata-anomaly-miner This tool parses log data and allows to define analysis pipelines for anomaly detection. It was designed to run the analysis wit

AECID 32 Nov 27, 2022
Universal data analysis tools for atmospheric sciences

U_analysis Universal data analysis tools for atmospheric sciences Script written in python 3. This file defines multiple functions that can be used fo

Luis Ackermann 1 Oct 10, 2021