# 547 search results for "boxplot"

## The 5th Tribe, Support Vector Machines and caret

October 15, 2015
By

by Joseph Rickert In his new book, The Master Algorithm, Pedro Domingos takes on the heroic task of explaining machine learning to a wide audience and classifies machine learning practitioners into 5 tribes*, each with its own fundamental approach to learning problems. To the 5th tribe, the analogizers, Pedro ascribes the Support Vector Machine (SVM) as it's master algorithm....

## A high quality plot

October 14, 2015
By

I’ll keep this post short and sweet. Here’s some code to get a really nice looking plot in R. It has a high pixel count to produce a high resolution output that can be used in a word document. Because of this, the size of everything in the plot (axes, points, text, axis labels, etc)

## Colouring a plot using a continuous variable in R

September 29, 2015
By

I often find myself coming back to this answer I gave on Stack Overflow in 2014. It shows how to colour a plot based on an independent continuous variable using the base graphics… Continue reading →

## Hypothesis-Driven Development Part V: Stop-Loss, Deflating Sharpes, and Out-of-Sample

September 24, 2015
By

This post will demonstrate a stop-loss rule inspired by Andrew Lo’s paper “when do stop-loss rules stop losses”? Furthermore, it … Continue reading →

## Convergence and Asymptotic Results

September 24, 2015
By
$\overline{X}_n\ \xrightarrow{\text{a.s.}}\ \mathbb{E}(X)$

Last week, in our mathematical statistics course, we’ve seen the law of large numbers (that was proven in the probability course), claiming that given a collection  of i.i.d. random variables, with To visualize that convergence, we can use > m=100 > mean_samples=function(n=10){ + X=matrix(rnorm(n*m),nrow=m,ncol=n) + return(apply(X,1,mean)) + } > B=matrix(NA,100,20) > for(i in 1:20){ + B=mean_samples(i*10) + } > colnames(B)=as.character(seq(10,200,by=10)) > boxplot(B) It is...

## Fitting a neural network in R; neuralnet package

September 23, 2015
By

Neural networks have always been one of the most fascinating machine learning model in my opinion, not only because of the fancy backpropagation algorithm, but also because of their complexity (think of deep learning with many hidden layers) and structure inspired by the brain. Neural networks have not always been popular, partly because they were,

## Predicting creditability using logistic regression in R: cross validating the classifier (part 2)

September 15, 2015
By

Now that we fitted the classifier and run some preliminary tests, in order to get a grasp at how our model is doing when predicting creditability we need to run some cross validation methods.Cross validation is a model evaluation method that does not u...

## Hypothesis Driven Development Part III: Monte Carlo In Asset Allocation Tests

September 10, 2015
By

This post will show how to use Monte Carlo to test for signal intelligence. Although I had rejected this strategy … Continue reading →

## Hypothesis-Driven Development Part II

September 8, 2015
By

This post will evaluate signals based on the rank regression hypotheses covered in the last post. The last time around, … Continue reading →

## Free R Help

September 3, 2015
By

Today I am giving away 10 sessions of free, online, one-on-one R help. My hope is to get a better understanding of how my readers use R, and the issues they face when working on their own projects. The sessions will be over the next two weeks, online and 30-60 minutes each. I just purchased Screenhero, The post