익명 사용자
로그인하지 않음
토론
기여
계정 만들기
로그인
IT 위키
검색
Data Science Cheat Sheet
편집하기
IT 위키
이름공간
문서
토론
더 보기
더 보기
문서 행위
읽기
편집
원본 편집
역사
경고:
로그인하지 않았습니다. 편집을 하면 IP 주소가 공개되게 됩니다.
로그인
하거나
계정을 생성하면
편집자가 사용자 이름으로 기록되고, 다른 장점도 있습니다.
스팸 방지 검사입니다. 이것을 입력하지
마세요
!
== Models == * '''Support Vector Machine (SVM)''': A supervised model that finds the optimal hyperplane for class separation, widely used in high-dimensional tasks like text classification (e.g., spam detection). ** '''''Advantage''''': Effective in high-dimensional spaces and robust to overfitting with the proper kernel. ** '''''Disadvantage''''': Computationally intensive on large datasets and sensitive to parameter tuning. * '''k-Nearest Neighbors (kNN)''': A non-parametric method that classifies based on nearest neighbors, often applied in recommendation systems and image recognition. ** '''''Advantage''''': Simple and intuitive, with no training phase, making it easy to implement. ** '''''Disadvantage''''': Computationally expensive at prediction time, especially with large datasets, and sensitive to irrelevant features. * '''Decision Tree''': A model that splits data into branches based on feature values, useful for interpretable applications like customer segmentation and medical diagnosis. ** '''''Advantage''''': Highly interpretable and handles both numerical and categorical data well. ** '''''Disadvantage''''': Prone to overfitting, especially with deep trees, and can be sensitive to small data changes. * '''Linear Regression''': A statistical technique that predicts a continuous outcome based on linear relationships, commonly used in financial forecasting and trend analysis. ** '''''Advantage''''': Simple and interpretable, with fast training for large datasets. ** '''''Disadvantage''''': Assumes a linear relationship, so it's unsuitable for complex, non-linear data. * '''Logistic Regression''': A classification model estimating the probability of a binary outcome, widely used in credit scoring and binary medical diagnostics. ** '''''Advantage''''': Interpretable with a clear probabilistic output, efficient for binary classification. ** '''''Disadvantage''''': Limited to linear boundaries, making it ineffective for complex relationships without transformations. * '''Naive Bayes''': A probabilistic classifier assuming feature independence, effective in text classification tasks like spam filtering due to its speed and simplicity. ** '''''Advantage''''': Fast and efficient, especially on large datasets with independence assumptions holding. ** '''''Disadvantage''''': Assumes feature independence, which may reduce accuracy if dependencies exist between features. == Confusion Matrix and F1 Score == '''[[Confusion Matrix]]''' {| class="wikitable" |- ! !!Predicted Positive!!Predicted Negative |- |'''Actual Positive'''||True Positive (TP)||False Negative (FN) |- |'''Actual Negative'''||False Positive (FP)||True Negative (TN) |} '''[[F1 Score]]''' = 2 * (Precision * Recall) / (Precision + Recall) * 2 * (Positive Predictive Value * True Positive Rate) / (Positive Predictive Value + True Positive Rate) * 2 * (TP) / (TP + FP + FN) == Key Evaluation Metrics == '''[[Recall (Data Science)|True Positive Rate (TPR), Sensitivity, Recall]]''' *TPR = Sensitivity = Recall = TP / (TP + FN) *Application: Measures the model's ability to correctly identify positive cases, useful in medical diagnostics to ensure true positives are detected. '''[[Precision (Data Science)|Precision (Positive Predictive Value)]]''' *Precision = TP / (TP + FP) *Application: Indicates the proportion of positive predictions that are correct, valuable in applications like spam filtering to minimize false alarms. '''[[Specificity (Data Science)|Specificity (True Negative Rate, TNR)]]''' *Specificity = TNR = TN / (TN + FP) *Application: Assesses the model's accuracy in identifying negative cases, crucial in fraud detection to avoid unnecessary scrutiny of legitimate transactions. '''[[False Positive Rate|False Positive Rate (FPR)]]''' *FPR = FP / (FP + TN) *Application: Reflects the rate of false alarms for negative cases, significant in security systems where false positives can lead to excessive interventions. '''Negative Predictive Value (NPV)''' *NPV = TN / (TN + FN) *Application: Shows the likelihood that a negative prediction is accurate, important in screening tests to reassure negative cases reliably. '''[[Accuracy (Data Science)|Accuracy]]''' *Accuracy = (TP + TN) / (TP + TN + FP + FN) *Application: Provides an overall measure of model correctness, often used as a baseline metric but less informative for imbalanced datasets. == Curves & Chart == '''[[Lift Curve]]''' * '''X-axis''': Percent of data (typically population percentile or cumulative population) * '''Y-axis''': Lift (ratio of model's performance vs. baseline) * '''Application''': Helps in evaluating the effectiveness of a model in prioritizing high-response cases, often used in marketing to identify segments likely to respond to promotions. '''[[Gain Chart]]''' * '''X-axis''': Percent of data (typically cumulative population) * '''Y-axis''': Cumulative gain (proportion of positives captured) * '''Application''': Illustrates the cumulative capture of positive responses at different cutoffs, useful in customer targeting to assess the efficiency of resource allocation. '''[[Cumulative Response Curve]]''' * '''X-axis''': Percent of data (cumulative population) * '''Y-axis''': Cumulative response (actual positives captured as cumulative total) * '''Application''': Evaluates model performance by showing how many true positives are captured as more of the population is included, applicable in direct marketing to optimize campaign reach. '''[[ROC Curve]]''' * '''X-axis''': False Positive Rate (FPR) * '''Y-axis''': True Positive Rate (TPR or Sensitivity) * '''Application''': Used to evaluate the trade-off between true positive and false positive rates at various thresholds, crucial in medical testing to balance sensitivity and specificity. '''[[Precision-Recall Curve]]''' * '''X-axis''': Recall (True Positive Rate) * '''Y-axis''': Precision (Positive Predictive Value) * '''Application''': Focuses on the balance between recall and precision, especially useful in cases of class imbalance, like fraud detection or medical diagnosis, where positive class accuracy is vital.
요약:
IT 위키에서의 모든 기여는 크리에이티브 커먼즈 저작자표시-비영리-동일조건변경허락 라이선스로 배포된다는 점을 유의해 주세요(자세한 내용에 대해서는
IT 위키:저작권
문서를 읽어주세요). 만약 여기에 동의하지 않는다면 문서를 저장하지 말아 주세요.
또한, 직접 작성했거나 퍼블릭 도메인과 같은 자유 문서에서 가져왔다는 것을 보증해야 합니다.
저작권이 있는 내용을 허가 없이 저장하지 마세요!
취소
편집 도움말
(새 창에서 열림)
둘러보기
둘러보기
대문
최근 바뀜
광고
위키 도구
위키 도구
특수 문서 목록
문서 도구
문서 도구
사용자 문서 도구
더 보기
여기를 가리키는 문서
가리키는 글의 최근 바뀜
문서 정보
문서 기록