HDDT is proposed as a technique for classifying data streams with unbalanced classes and concept drift. HDDT uses Hellinger distance as the splitting criteria in decision trees, which performs better than standard decision trees on unbalanced data. An ensemble of HDDTs is used to handle concept drift. Experimental results on several datasets show that HDDT without instance propagation between batches outperforms state-of-the-art techniques that use propagation, and HDDT ensembles have higher accuracy than single classifiers.