통일된 DataScience 폴더 구조 제공 및 가상환경 작업의 부담감 해소

Related tags

Deep LearningLucas
Overview


Lucas

Hits


coded by linux shell

목차


Patch Note 📜


Team member

Contributors/People

ympark gbhwang cbchun
https://github.com/pym7857 https://github.com/gbhwang https://github.com/bermmie1000
  • You can see team member and github profile
  • You should probably find team member's lastest project



Requirements

  • python 3.xx



Mac버전 CookieCutter (autoenv)

🚫 주의
$> brew install autoenv 로 다운로드 받아서 실행시키면 터미널 고장납니다.
반드시 autoenv Github 에서 git clone 으로 다운받아 주세요. (현재 시점 21.3.24)

⚠️ mac버전만 소개합니다.

1. How to Install autoenv

$ git clone git://github.com/inishchith/autoenv.git ~/.autoenv

2.폴더 진입 시, activate 구현하기

$ echo 'source ~/.autoenv/activate.sh' >> ~/.zshrc
$ source ~/.zshrc

🔔 하단의.env파일은 현재 repo의 cookiecutter에서 자동으로 생성해줍니다. (스킵)

# .env 파일
echo "HELLO autoenv"
{
    source .dev-venv/bin/activate
    echo "virtual env is successfully activated!"
} ||
{
    echo "[virtual env start] is failed!"
}

.env파일 설정 후 첫 폴더 진입시 .env파일을 신뢰하고 실행할지 않을 지에 대한 동의가 나타납니다. autoenv 이 부분은 .env파일이 악의적으로 변경되었을때 사용자에게 알리기 위해서 있기 때문에 즐거운 마음으로 Y를 눌러줍시다.
이제 정상적으로 가상환경이 activate된 것을 확인할 수 있습니다.

3.폴더 탈출 시, deactivate 구현하기

$> vi ~/.zshrc

마지막줄에 다음의 명령어를 추가해줍니다.

export AUTOENV_ENABLE_LEAVE='"enabled"' 

🔔 하단의.env.leave파일은 현재 repo의 cookiecutter에서 자동으로 생성해줍니다. (스킵)

# .env.leave 파일
echo "BYEBYE"
{
    deactivate
    echo "virtual env is successfully deactivated!"
} ||
{
    echo "[virtual env quit] is failed!"
}

.env.leave파일 설정 후 해당 폴더에서 나가면
정상적으로 가상환경이 deactivate 되는 것을 확인할 수 있습니다.

4.Alias 설정하기

echo 'alias cookie="bash [각자 컴퓨터의 상대경로/cookie_cutter_project_dir.sh]"' >> ~/.zshrc
ex) echo 'alias cookie="bash /Users/gbhwang/Desktop/Project/Test/Lucas/mac/cookie_cutter_project_dir.sh"' >> ~/.zshrc

맥 파일경로 확인법을 참고하여
각자 mac폴더안의 cookie_cutter_project_dir.sh 파일의 경로를 확인하여 zshrc에 넣어주시면 됩니다.

이렇게 하면 cookie 명령어 만으로 간단하게 스크립트를 실행시킬 수 있게 됩니다.
위와 같이 설정하면 cookie [프로젝트 생성할 경로] [프로젝트 이름] 명령어로 프로젝트를 생성할 수 있게 됩니다.

5.How to Use

$> cd "where-you-want"
$> git clone https://github.com/LS-ELLO/Lucas.git
$> cd Lucas
$> cd mac

$> cookie [where-you-want] [your-project-name]
ex) $> cookie . test111



Windows버전 CookieCutter (ps-autoenv)

도움 주신 규본님 감사합니다.
ps-autoenv를 사용합니다.

1.How to install ps-autoenv

Powershell 실행 (관리자 권한 실행)

PS> Install-Module ps-autoenv
PS> Add-Content $PROFILE @("`n", "import-module ps-autoenv")

2.Alias 설정하기 (git-bash)

참조

  1. C:/Program Files/Git/etc/profile.d/aliases.sh 파일을 관리자 권한으로 Text Editor에 실행시킵니다.

  2. 다음의 명령어를 추가합니다.
    alias cookie='bash cookie_cutter_project_dir.sh의 상대경로'
    ex) alias cookie='bash D:/Lucas/windows/cookie_cutter_project_dir.sh'

    (aliases.sh)

    # Some good standards, which are not used if the user
    # creates his/her own .bashrc/.bash_profile
    
    # --show-control-chars: help showing Korean or accented characters
    alias ls='ls -F --color=auto --show-control-chars'
    alias ll='ls -l'
    alias cookie='bash [where-your-cookie_cutter_project_dir.sh]'
    
    case "$TERM" in
    ...

3.How to Use

Git Bash 실행

bash> cd "where-this-repo-downloaded"
bash> cd windows
bash> cookie [where-you-want] [your-project-name]
ex) cookie . 1bot

Powershell 실행

PS> Import-Module ps-autoenv
PS> cd "where-your-cookiecutter-project"
ex. PS> cd "C:\Users\ympark4\Documents\1bot"
PS> press 'Y'
🚫 PSSecurityException 오류 발생할때

https://extbrain.tistory.com/118 를 참조해서 해결주세요.



The resulting directory structure

The directory structure of your new project looks like this:

├── LICENSE
├── Makefile
├── README.md          ← The top-level README for developers using this project.
├── data
│   ├── external       ← Data from third party sources.
│   ├── interim        ← Intermediate data that has been transformed.
│   ├── processed      ← The final, canonical data sets for modeling.
│   └── raw            ← The original, immutable data dump.
├── docs               ← A default Sphinx project; see sphinx-doc.org for details
├── models             ← Trained and serialized models, model predictions, or model summaries
├── notebooks          ← Jupyter notebooks. Naming convention is a number (for ordering), the creator's initials, and a short `-` delimited description, e.g. `1.0-jqp-initial-data-exploration`.
├── references         ← Data dictionaries, manuals, and all other explanatory materials.
├── reports            ← Generated analysis as HTML, PDF, LaTeX, etc.
│   └── figures        ← Generated graphics and figures to be used in reporting
├── requirements.txt   ← The requirements file for reproducing the analysis environment, e.g. generated with `pip freeze > requirements.txt`
├── setup.py           ← makes project pip installable (pip install -e .) so src can be imported
├── src                ← Source code for use in this project.
│   ├── __init__.py  
│   ├── dataread      
│   │   └── __init__.py
│   │   └── example.py
│   │
│   ├── features       
│   │   └── __init__.py
│   │   └── example.py
│   │
│   ├── models     
│   │   └── __init__.py
│   │   └── example.py
│   │
│   ├── visualization    
│   │   └── __init__.py
│   │   └── example.py
├── App               
│   ├── android       
│   ├── ios           
│   ├── lib            
│   │   └── models
│   │   └── main.dart
│
└── .gitignore        



Owner
ello
ello
Amazing-Python-Scripts - 🚀 Curated collection of Amazing Python scripts from Basics to Advance with automation task scripts.

📑 Introduction A curated collection of Amazing Python scripts from Basics to Advance with automation task scripts. This is your Personal space to fin

Avinash Ranjan 1.1k Dec 29, 2022
Inference pipeline for our participation in the FeTA challenge 2021.

feta-inference Inference pipeline for our participation in the FeTA challenge 2021. Team name: TRABIT Installation Download the two folders in https:/

Lucas Fidon 2 Apr 13, 2022
Serve TensorFlow ML models with TF-Serving and then create a Streamlit UI to use them

TensorFlow Serving + Streamlit! ✨ 🖼️ Serve TensorFlow ML models with TF-Serving and then create a Streamlit UI to use them! This is a pretty simple S

Álvaro Bartolomé 18 Jan 07, 2023
An end-to-end implementation of intent prediction with Metaflow and other cool tools

You Don't Need a Bigger Boat An end-to-end (Metaflow-based) implementation of an intent prediction flow for kids who can't MLOps good and wanna learn

Jacopo Tagliabue 614 Dec 31, 2022
DanceTrack: Multiple Object Tracking in Uniform Appearance and Diverse Motion

DanceTrack DanceTrack is a benchmark for tracking multiple objects in uniform appearance and diverse motion. DanceTrack provides box and identity anno

260 Dec 28, 2022
An addernet CUDA version

Training addernet accelerated by CUDA Usage cd adder_cuda python setup.py install cd .. python main.py Environment pytorch 1.10.0 CUDA 11.3 benchmark

LingXY 4 Jun 20, 2022
QRec: A Python Framework for quick implementation of recommender systems (TensorFlow Based)

Introduction QRec is a Python framework for recommender systems (Supported by Python 3.7.4 and Tensorflow 1.14+) in which a number of influential and

Yu 1.4k Jan 01, 2023
Exploring Cross-Image Pixel Contrast for Semantic Segmentation

Exploring Cross-Image Pixel Contrast for Semantic Segmentation Exploring Cross-Image Pixel Contrast for Semantic Segmentation, Wenguan Wang, Tianfei Z

Tianfei Zhou 510 Jan 02, 2023
Scripts and outputs related to the paper Prediction of Adverse Biological Effects of Chemicals Using Knowledge Graph Embeddings.

Knowledge Graph Embeddings and Chemical Effect Prediction, 2020. Scripts and outputs related to the paper Prediction of Adverse Biological Effects of

Knowledge Graphs at the Norwegian Institute for Water Research 1 Nov 01, 2021
Pytorch implementation of the paper Time-series Generative Adversarial Networks

TimeGAN-pytorch Pytorch implementation of the paper Time-series Generative Adversarial Networks presented at NeurIPS'19. Jinsung Yoon, Daniel Jarrett

Zhiwei ZHANG 21 Nov 24, 2022
Python Blood Vessel Topology Analysis

Python Blood Vessel Topology Analysis This repository is not being updated anymore. The new version of PyVesTo is called PyVaNe and is available at ht

6 Nov 15, 2022
A PyTorch-based Semi-Supervised Learning (SSL) Codebase for Pixel-wise (Pixel) Vision Tasks

PixelSSL is a PyTorch-based semi-supervised learning (SSL) codebase for pixel-wise (Pixel) vision tasks. The purpose of this project is to promote the

Zhanghan Ke 255 Dec 11, 2022
Neural style transfer as a class in PyTorch

pt-styletransfer Neural style transfer as a class in PyTorch Based on: https://github.com/alexis-jacq/Pytorch-Tutorials Adds: StyleTransferNet as a cl

Tyler Kvochick 31 Jun 27, 2022
Try out deep learning models online on Google Colab

Try out deep learning models online on Google Colab

Erdene-Ochir Tuguldur 1.5k Dec 27, 2022
Simulation-based performance analysis of server-less Blockchain-enabled Federated Learning

Blockchain-enabled Server-less Federated Learning Repository containing the files used to reproduce the results of the publication "Blockchain-enabled

Francesc Wilhelmi 9 Sep 27, 2022
[ICLR 2021] "CPT: Efficient Deep Neural Network Training via Cyclic Precision" by Yonggan Fu, Han Guo, Meng Li, Xin Yang, Yining Ding, Vikas Chandra, Yingyan Lin

CPT: Efficient Deep Neural Network Training via Cyclic Precision Yonggan Fu, Han Guo, Meng Li, Xin Yang, Yining Ding, Vikas Chandra, Yingyan Lin Accep

26 Oct 25, 2022
Code for the paper BERT might be Overkill: A Tiny but Effective Biomedical Entity Linker based on Residual Convolutional Neural Networks

Biomedical Entity Linking This repo provides the code for the paper BERT might be Overkill: A Tiny but Effective Biomedical Entity Linker based on Res

Tuan Manh Lai 24 Oct 24, 2022
PyTorch reimplementation of REALM and ORQA

PyTorch reimplementation of REALM and ORQA

Li-Huai (Allan) Lin 17 Aug 20, 2022
Code release for DS-NeRF (Depth-supervised Neural Radiance Fields)

Depth-supervised NeRF: Fewer Views and Faster Training for Free Project | Paper | YouTube Pytorch implementation of our method for learning neural rad

524 Jan 08, 2023
Ground truth data for the Optical Character Recognition of Historical Classical Commentaries.

OCR Ground Truth for Historical Commentaries The dataset OCR ground truth for historical commentaries (GT4HistComment) was created from the public dom

Ajax Multi-Commentary 3 Sep 08, 2022