决策树学习

决策树

决策树基本算法

划分选择

‌‌‌‌

‌‌‌‌  决策树最关键的就是如何选择划分属性。一般而言,随着划分过程不断进行,我们希望决策树的分支节点所包含的样本尽可能属于同一类别,即结点的"纯度"(purity)越来越高

信息增益

‌‌‌‌  "信息熵"(information entropy)是度量样本集合纯度最常用的一种指标。假定当前样本集合DDD中第kkk类样本所占比例为pk(k=1,2,⋯ ,∣Y∣)p_k\left(k=1,2,\cdots, \left|\mathcal{Y}\right|\right)pk(k=1,2,⋯,∣Y∣), 则DDD的信息熵定义为
Ent(D)=−∑k=1∣Y∣pklog⁡2pk \mathrm{Ent}\left( D \right) = -\sum_{k=1}^{\left| \mathcal{Y} \right| }p_{k} \log_{2} p_{k} Ent(D)=−k=1∑∣Y∣pklog2pk
Ent(D)\mathrm{Ent}\left( D \right)Ent(D)的值越小,则DDD的纯度越高 (可以证明均匀分布的时候最大,其中一个为1的时候最小)

python 复制代码
def calc_entropy(df: pd.DataFrame, label_col: str) -> float:
    if df.empty:
        return 0.0
    counts = df[label_col].value_counts(normalize=True)
    # -(|C_k| / |D|) * log2(|C_k| / |D|)
    return -np.sum(counts * np.log2(counts))

‌‌‌‌  假定离散属性aaa有VVV个可能的取值{a1,a2,⋯ ,aV}\left\{ a^1, a^2,\cdots,a^V \right\}{a1,a2,⋯,aV}, 若使用aaa来对样本集DDD进行划分,则会产生VVV个分支结点,其中第vvv个分支节点包含了DDD中所有在属性aaa上取值为ava^vav的样本,记为DvD^vDv.我们可以根据信息熵公式计算出DvD^vDv的信息熵,在考虑到不同分支节点所包含的样本数不同,给分支节点赋予权重∣Dv∣∣D∣\frac{\left| D^v \right|}{\left| D \right|}∣D∣∣Dv∣,即样本数越多的分支节点的影响越大,于是可以算出用属性aaa对样本集DDD进行划分所获得的"信息增益"(information gain)
Gain(D,a)=Ent(D)−∑v=1V∣Dv∣∣D∣Ent(Dv) \mathrm{Gain}\left( D, a \right) =\mathrm{Ent}\left( D \right) - \sum_{v=1}^{V} \frac{\left| D^v \right| }{\left| D \right| } \mathrm{Ent}\left( D^v \right) Gain(D,a)=Ent(D)−v=1∑V∣D∣∣Dv∣Ent(Dv)

python 复制代码
def calc_feature_entropy(df: pd.DataFrame, feature_name: str, label_col: str) -> float:
    if df.empty:
        return 0.0

    counts = df[feature_name].value_counts(normalize=True)
    # (|D_i| / |D|) * entropy(D_i)
    return sum(
        p * calc_entropy(df[df[feature_name] == feature], label_col)
        for feature, p in counts.items()
    )

‌‌‌‌  一般而言,信息增益越大,则意味着使用属性aaa进行划分所获得的"纯度提升"越大。因此,我们可以用信息增益来进行决策树的划分属性选择,即a=arg⁡max⁡a∈AGain(D,a)a = \arg \max_{a\in A} \mathrm{Gain}\left( D, a \right)a=argmaxa∈AGain(D,a).著名的ID3决策树学习算法就是以信息增益为准则来选择划分属性。

增益率

实际上,信息增益准则对可取值数目较多的属性有所偏好,为减少这种偏好可能带来的不利影响,著名的C4.5决策树算法不直接使用信息增益,而是使用"增益率"(gain ratio)来选择最优化分属性。增益率定义为:
Gain_ratio(D,a)=Gain(D,a)IV(a) \mathrm{Gain\_ratio} \left( D, a \right) = \frac{\mathrm{Gain}\left( D, a \right) }{\mathrm{IV}\left( a \right) } Gain_ratio(D,a)=IV(a)Gain(D,a)

其中
IV(a)=−∑v=1V∣Dv∣∣D∣log⁡2∣Dv∣∣D∣ \mathrm{IV}\left( a \right) =- \sum_{v=1}^{V}\frac{\left| D^{v} \right| }{\left| D \right| } \log_{2}\frac{\left| D^{v} \right| }{\left| D \right| } IV(a)=−v=1∑V∣D∣∣Dv∣log2∣D∣∣Dv∣

(相当于对特征那一列算信息熵)

称为属性aaa的"固有值"(intrinsic value)。属性aaa的的可能取值数目越多(即VVV越大),则IV(a)\mathrm{IV}\left( a \right)IV(a)的值通常会越大。

‌‌‌‌  需要注意的是,增益率准则对可取值数目较少的属性有所偏好,因此C4.5蒜贩并不是直接选择增益率最大的候选划分属性,而是使用了一个启发式:先从候选划分属性中找出信息增益高于平均水平的属性,再从中选择增益率最高的。

基尼指数

‌‌‌‌  CART决策树使用"基尼指数"(Gini index)来选择划分属性。数据集DDD的纯度可用基尼值来度量:
Gini(D)=∑k=1∣Y∣∑k′≠kpkpk′=1−∑k=1∣Y∣pk2 \begin{aligned} \mathrm{Gini}\left( D \right) &= \sum_{k=1}^{\left| \mathcal{Y} \right| }\sum_{k^{\prime}\neq k} p_{k} p_{k^{\prime}}\\ &= 1 - \sum_{k=1}^{\left| \mathcal{Y} \right| }p_{k}^{2} \end{aligned} Gini(D)=k=1∑∣Y∣k′=k∑pkpk′=1−k=1∑∣Y∣pk2

‌‌‌‌  直观来说,Gini(D)\mathrm{Gini}\left( D \right)Gini(D)反映了从数据集DDD中随机抽取两个样本,其类别标记不一致的概率。因此,Gini(D)\mathrm{Gini}\left( D \right)Gini(D)越小,则数据集DDD的纯度越高。

‌‌‌‌  属性aaa的基尼指数定义为
Gini_index(D,a)=∑v=1V∣Dv∣∣D∣Gini(Dv) \mathrm{Gini\index}\left( D, a \right) = \sum{v=1}^{V} \frac{\left| D^v \right| }{\left| D \right| }\mathrm{Gini}\left( D^v \right) Gini_index(D,a)=v=1∑V∣D∣∣Dv∣Gini(Dv)

于是,我们在候选属性集合中,选择那个使得划分后基尼指数最小的属性作为最优划分属性,即a∗=arg⁡min⁡a∈AGini_index(D,a)a_{*} = \arg\min_{a\in A} \mathrm{Gini\_index}\left( D,a \right)a∗=argmina∈AGini_index(D,a)

连续值

‌‌‌‌  由于连续属性的可取值数目不在有限,因此,不能直接根据连续属性的可取值来对结点进行划分。此时连续属性离散化技术可派上用场。最简单的策略是采用二分法对连续属性进行处理,这正是C4.5决策树算法中采用的机制

‌‌‌‌  给定样本集DDD和连续属性aaa,假定aaa在DDD熵出现的nnn个不同取值,将这些值从小到大进行排序,记为{a1,a2,⋯ ,an}\left\{ a^1, a^2,\cdots, a^n \right\}{a1,a2,⋯,an}。基于划分点ttt可将DDD分为子集Dt−D_{t}^{-}Dt−和Dt+D_{t}^{+}Dt+,其中Dt−D_{t}^{-}Dt−包含那些属性aaa上取值不大于ttt的样本,而Dt+D_{t}^{+}Dt+则包含那些在属性aaa上取值大于ttt的样本。显然,对相邻的属性取值aia^iai与ai+1a^{i+1}ai+1来说,ttt在区间[ai,ai+1)\left[a^{i}, a^{i+1} \right)[ai,ai+1)中取任意值所产生的划分结果相同。因此,对连续属性aaa,我们可考察包含n−1n-1n−1个元素的候选划分点集合
Ta={ai+ai+12:1≤i≤n−1} T_{a} = \left\{ \frac{a^{i} + a^{i+1}}{2}:1 \le i \le n -1 \right\} Ta={2ai+ai+1:1≤i≤n−1}

因此
Gain(D,a)=max⁡t∈TaGain(D,a,t)=max⁡t∈TaEnt(D)−∑λ∈{−,+}∣Dtλ∣∣D∣Ent(Dtλ) \begin{aligned} \mathrm{Gain}\left( D, a \right) &= \max\limits_{t \in T_{a}} \mathrm{Gain}\left( D, a, t \right) \\ &= \max\limits_{t \in T_{a}} \mathrm{Ent}\left( D \right) - \sum_{\lambda \in \left\{ -, + \right\} }\frac{\left| D_{t}^{\lambda} \right| }{\left| D \right| } \mathrm{Ent} \left( D_{t}^{\lambda} \right) \end{aligned} Gain(D,a)=t∈TamaxGain(D,a,t)=t∈TamaxEnt(D)−λ∈{−,+}∑∣D∣ Dtλ Ent(Dtλ)

固有值
IV(a,t)=−∑λ∈{−,+}∣Dtλ∣∣D∣log⁡2∣Dtλ∣∣D∣ \mathrm{IV}\left( a,t \right) = - \sum_{\lambda \in \left\{ -, + \right\} }\frac{\left| D_{t}^{\lambda} \right| }{\left| D \right| } \log_{2}\frac{\left| D_{t}^{\lambda} \right| }{\left| D \right| } IV(a,t)=−λ∈{−,+}∑∣D∣ Dtλ log2∣D∣ Dtλ

信息增益率
Gain_ratio(D,a)=max⁡t∈TaGain_ratio(D,a,t)=max⁡t∈TaGain(D,a,t)IV(a,t) \begin{aligned} \mathrm{Gain\ratio} \left( D, a \right) &= \max\limits{t \in T_{a}} \mathrm{Gain\ratio}\left( D, a, t \right) \\ &= \max\limits{t \in T_{a}} \frac{\mathrm{Gain}\left( D, a, t \right)}{\mathrm{IV}\left( a, t \right) } \end{aligned} Gain_ratio(D,a)=t∈TamaxGain_ratio(D,a,t)=t∈TamaxIV(a,t)Gain(D,a,t)

基尼指数
Gini_index(D,a)=min⁡t∈Ta∑λ∈{−,+}∣Dtλ∣∣D∣Gini(Dtλ) \begin{aligned} \mathrm{Gini\index}\left( D,a \right) = \min{t \in T_{a}} \sum_{\lambda \in \left\{ -, + \right\} }\frac{\left| D_{t}^{\lambda} \right| }{\left| D \right| } \mathrm{Gini}\left( D_{t}^{\lambda} \right) \end{aligned} Gini_index(D,a)=t∈Taminλ∈{−,+}∑∣D∣ Dtλ Gini(Dtλ)

缺失值

以下只写离散的,连续的可以当作有阈值后分成两个类,然后VVV只有2个取值

‌‌‌‌  我们需要解决两个问题:(1)如何在属性值缺失的情况下进行划分属性选择?(2)给定划分属性,若样本在该属性上的值缺失,如何对样本进行划分?

给定训练集DDD和属性aaa,令D~\tilde{D}D~表示DDD中在属性aaa上没有缺失值的样本子集。对问题(1),显然我们仅可根据D~\tilde{D}D~来判断属性aaa的优劣。假定属性aaa有VVV个可取值{a1,a2,⋯ ,aV}\left\{ a^1, a^2, \cdots, a^V \right\}{a1,a2,⋯,aV},令D~v\tilde{D}^{v}D~v表示D~\tilde{D}D~中在属性aaa上取值为ava^vav的样本子集,D~k\tilde{D}{k}D~k表示D~\tilde{D}D~中属于第kkk类(k=1,2,⋯ ,∣Y∣k=1,2,\cdots, \left| \mathcal{Y} \right|k=1,2,⋯,∣Y∣)的样本子集,显然D~=∪k=1∣Y∣D~k=∪v=1VD~v\tilde{D} = \cup{k=1}^{\left| \mathcal{Y} \right|} \tilde{D}{k}= \cup{v=1}^{V}\tilde{D}^vD~=∪k=1∣Y∣D~k=∪v=1VD~v。假定我们为每个样本x\mathbf{x}x赋予一个权重wxw_{\mathbf{x}}wx(可以考虑全1),并定义
ρ=∑x∈D~wx∑x∈Dwx \rho = \frac{\sum_{\mathbf{x} \in \tilde{D}} w_{\mathbf{x}}}{\sum_{x\in D} w_{\mathbf{x}}} ρ=∑x∈Dwx∑x∈D~wx
p~k=∑x∈D~kwx∑x∈D~wx(1≤k≤∣Y∣) \tilde{p}{k} = \frac{\sum{x \in \tilde{D}{k}} w{\mathbf{x}}}{\sum_{\mathbf{x} \in \tilde{D}}w_{\mathbf{x}}} \left( 1 \le k \le \left| \mathcal{Y} \right| \right) p~k=∑x∈D~wx∑x∈D~kwx(1≤k≤∣Y∣)
r~v=∑x∈D~vwx∑x∈D~wx(1≤v≤V) \tilde{r}{v} = \frac{\sum{\mathbf{x} \in \tilde{D}^{v}} w_{\mathbf{x}}}{\sum_{\mathbf{x} \in \tilde{D}}w_{\mathbf{x}}} \left( 1 \le v \le V \right) r~v=∑x∈D~wx∑x∈D~vwx(1≤v≤V)
p~vk=∑x∈D~kvwx∑x∈D~vwx(1≤v≤V) \tilde{p}{v}^{k} = \frac{\sum{\mathbf{x} \in \tilde{D}{k}^{v}} w{\mathbf{x}}}{\sum_{\mathbf{x} \in \tilde{D}^{v}}w_{\mathbf{x}}} \left( 1 \le v \le V \right) p~vk=∑x∈D~vwx∑x∈D~kvwx(1≤v≤V)

直观理解,ρ\rhoρ是无缺失值样本所占比例
p~k\tilde{p}{k}p~k表示无缺失值样本中第kkk类所占的比例
r~v\tilde{r}
{v}r~v表示无缺失值样本中在属性aaa上取值ava^vav的样本所占比例
p~vk\tilde{p}_{v}^{k}p~vk表示属性aaa上取值ava^vav的样本中第kkk类所占的比例

Ent(D~)=−∑k=1∣Y∣p~klog⁡2p~k \mathrm{Ent}\left( \tilde{D} \right)= -\sum_{k=1}^{\left| \mathcal{Y} \right| }\tilde{p}{k}\log{2}\tilde{p}{k} Ent(D~)=−k=1∑∣Y∣p~klog2p~k
Ent(D~v)=−∑k=1∣Y∣p~vklog⁡2p~vk \mathrm{Ent}\left( \tilde{D}^{v} \right)= -\sum
{k=1}^{\left| \mathcal{Y} \right| }\tilde{p}{v}^{k}\log{2}\tilde{p}_{v}^{k} Ent(D~v)=−k=1∑∣Y∣p~vklog2p~vk

python3 复制代码
def entropy(categories: pd.Series, weight: np.ndarray = None) -> float:
    if categories.empty:
        return 0.0

    if weight is None:
        counts = categories.value_counts(normalize=True)
        return -np.sum(counts * np.log2(counts))

    # 将类别转换为整数标签并计算加权频数
    labels, _ = pd.factorize(categories, sort=False)
    counts = np.bincount(labels, weights=weight, minlength=0)

    # 计算总权重和概率
    total_weight = counts.sum()
    if total_weight <= 0:
        return 0.0

    probabilities = counts / total_weight

    # 仅处理非零概率避免log(0)警告
    non_zero_probs = probabilities[probabilities > 0]
    return -np.sum(non_zero_probs * np.log2(non_zero_probs))

Gain(D,a)=ρ×Gain(D~,a)=ρ(Ent(D~)−∑v=1Vr~kEnt(D~v)) \begin{aligned} \mathrm{Gain}\left( D, a \right) &= \rho \times \mathrm{Gain}\left( \tilde{D}, a \right) \\ &= \rho \left( \mathrm{Ent}\left( \tilde{D} \right) - \sum_{v=1}^{V} \tilde{r}{k} \mathrm{Ent}\left( \tilde{D}^{v} \right) \right) \\ \end{aligned} Gain(D,a)=ρ×Gain(D~,a)=ρ(Ent(D~)−v=1∑Vr~kEnt(D~v))
IV(D)=ρIV(D~)=ρ(−∑v=1Vr~vlog⁡2r~v) \mathrm{IV}\left( D \right)=\rho IV\left( \tilde{D} \right) =\rho \left( -\sum
{v=1}^{V}\tilde{r}{v}\log{2}\tilde{r}_{v} \right) IV(D)=ρIV(D~)=ρ(−v=1∑Vr~vlog2r~v)

python 复制代码
# df_feature不包含na,weight_feature也不包含na
feature_values = df_feature[feature].unique()
# r_k * entropy(\tilde{D}^v)
feature_entropy = 0.0
for value in feature_values:
	mask = df_feature[feature] == value
	value_weight = weight_feature[mask]
	value_entropy = entropy(df_feature.loc[mask, label_col], value_weight)
	rk = value_weight.sum() / weight_feature_sum
	feature_entropy += rk * value_entropy
info_gain = rho * (total_entropy - feature_entropy)
iv = rho * entropy(df_feature[feature], weight_feature)  # 固有值

Gain_ratio(D,a)=Gain(D,a)IV(a) \mathrm{Gain\ratio} \left( D, a \right) = \frac{\mathrm{Gain}\left( D, a \right) }{\mathrm{IV}\left( a \right) } Gain_ratio(D,a)=IV(a)Gain(D,a)
Gini(D~)=1−∑k=1∣Y∣p~k \mathrm{Gini}\left( \tilde{D} \right) =1 - \sum
{k=1}^{\left| \mathcal{Y} \right| }\tilde{p}{k} Gini(D~)=1−k=1∑∣Y∣p~k
Gini(D~v)=1−∑k=1∣Y∣p~vk \mathrm{Gini}\left( \tilde{D}^v \right) =1 - \sum
{k=1}^{\left| \mathcal{Y} \right| }\tilde{p}_{v}^{k} Gini(D~v)=1−k=1∑∣Y∣p~vk

python 复制代码
def gini(categories: pd.Series, weight: np.ndarray = None) -> float:
    if categories.empty:
        return 0.0
    if weight is None:
        counts = categories.value_counts(normalize=True)
        # 1 - sum(p_k^2)
        return 1 - np.sum(counts**2)

    # 将类别转换为整数标签并计算加权频数
    labels, _ = pd.factorize(categories, sort=False)
    counts = np.bincount(labels, weights=weight, minlength=0)

    # 计算总权重和概率
    total_weight = counts.sum()
    if total_weight <= 0:
        return 0.0

    probabilities = counts / total_weight
    return 1 - np.sum(probabilities**2)

Gini_index(D,a)=ρ∑v=1Vr~vGini(D~v) \begin{aligned} \mathrm{Gini\index}\left( D, a \right) = \rho \sum{v=1}^{V} \tilde{r}_{v}\mathrm{Gini}\left( \tilde{D}^{v} \right) \end{aligned} Gini_index(D,a)=ρv=1∑Vr~vGini(D~v)

‌‌‌‌  对问题(2),若样本x\mathbf{x}x在划分属性aaa上的取值已知,则将x\mathbf{x}x划入与其取值对应的子节点,且样本全职在子结点中保持为wxw_{\mathbf{x}}wx。若样本x\mathbf{x}x在划分属性aaa上的取值未知,则将x\mathbf{x}x同时划入所有的子结点,且样本权值在与属性值ava^vav对应的子结点中调整为r~vwx\tilde{r}{v} w{\mathbf{x}}r~vwx;直观来看,这就是让同一个样本以不同的概率划入到不同的子结点中去

‌‌‌‌  C4.5算法使用了上述解决方案

推理的时候也是按权重得到子结点,相同类别权重相加,然后找到权重最高的类别

Cart

cart算法会得到一颗二叉树

分类

假定离散属性aaa有VVV个可能的取值{a1,a2,⋯ ,aV}\left\{ a^1, a^2,\cdots,a^V \right\}{a1,a2,⋯,aV},可以根据aia^iai将DDD分为两类,D1D^1D1是取值为aia^iai,另一类是D2=D−D1D^2=D-D^1D2=D−D1

对于连续的天然就是2个类
Gini_index(D,a)=min⁡ai∈{a1,a2,⋯ ,aV}ρ∑v=12r~vGini(D~v) \begin{aligned} \mathrm{Gini\index}\left( D, a \right) = \min{a^i\in \left\{ a^1, a^2,\cdots,a^V \right\} }\rho \sum_{v=1}^{2} \tilde{r}_{v}\mathrm{Gini}\left( \tilde{D}^{v} \right) \end{aligned} Gini_index(D,a)=ai∈{a1,a2,⋯,aV}minρv=1∑2r~vGini(D~v)

回归

也是分成两类,不妨假设sss是切分点
min⁡a,sρ[min⁡c1∑x∈R1wi(yi−c1)2+min⁡c2∑x∈R2wi(yi−c1)2] \min_{a,s}\rho \left[ \min_{c_{1}} \sum_{x \in R_{1}}w_{i}\left( y_{i} - c_{1} \right)^2 + \min_{c_{2}} \sum_{x \in R_{2}}w_{i}\left( y_{i} - c_{1} \right)^2\right] a,sminρ[c1minx∈R1∑wi(yi−c1)2+c2minx∈R2∑wi(yi−c1)2]

其中R1,R2R_{1}, R_{2}R1,R2均不包含缺失值

设W=diag(w)\mathbf{W} = \mathrm{diag}\left( \mathbf{w} \right)W=diag(w)
min⁡c∑wi(yi−c)2=(y−c1)W(y−c1) \min_{c}\sum w_{i} \left( y_{i}-c \right)^2 = \left( \mathbf{y}-c \mathbf{1} \right) \mathbf{W}\left( \mathbf{y} - c \mathbf{1} \right) cmin∑wi(yi−c)2=(y−c1)W(y−c1)
∇L=21TW(y−c1)=0⇒c=1TWy1TW1=∑wiyi∑wi \nabla L = 2\mathbf{1}^T\mathbf{W}\left( \mathbf{y} - c \mathbf{1} \right) =0 \Rightarrow c = \frac{\mathbf{1}^T \mathbf{W}\mathbf{y}}{\mathbf{1}^T \mathbf{W}\mathbf{1}}=\frac{\sum w_{i}y_{i}}{\sum w_{i}} ∇L=21TW(y−c1)=0⇒c=1TW11TWy=∑wi∑wiyi

然后每个结点的值就是∑wiyi∑wi\frac{\sum w_{i}y_{i}}{\sum w_{i}}∑wi∑wiyi,也就是加权平均

剪枝

‌‌‌‌  剪枝是决策树学习算法对付"过拟合"的主要手段。在决策树学习中,为了尽可能正确分类训练样本,结点划分过程不断重复,有时会造成决策树分支过多,这时就可能因训练样本学得"太好"了,以至于把训练集自身的一些特点当作所有数据都具有的一般性质而导致过拟合。因此,可通过主动去掉一些分支来降低过拟合的风险

预剪枝

先计算一下验证集的acc,然后计算一下划分后的验证集acc;如果划分前的acc>=划分后的,那就不划分了。

一种方便的做法就是,直接算一下,然后先添加子节点算一遍acc,然后清空

后剪枝

基于acc

对于每一个叶子的父结点(即该节点的孩子都是叶结点),先计算一下划分后的验证集acc;然后计算一下裁剪掉acc,如果裁剪后acc>裁剪前的acc就裁剪

然后重复这个过程

做法类似预剪枝,相当于后序遍历,遍历完孩子后

Minimal Cost-Complexity Pruning

下面小写字母表示结点,例如ttt

大写字母表示树,例如TTT
TtT_{t}Tt表示以ttt为根节点的树
∣Tt∣\left| T_{t} \right|∣Tt∣表示TtT_tTt的叶子节点个数
leaf(T)\mathrm{leaf}\left( T \right)leaf(T)表示TTT的叶子节点

损失函数
Cα(T)=C(T)+α∣T∣=(∑t∈leaf(T)p(t)R(t))+α∣T∣ C_{\alpha}\left( T \right) = C\left( T \right) + \alpha \left| T \right| =\left( \sum_{t \in \mathrm{leaf(T)}} p\left( t \right) R\left( t \right) \right) + \alpha \left| T \right| Cα(T)=C(T)+α∣T∣= t∈leaf(T)∑p(t)R(t) +α∣T∣

其中p(t)p\left( t \right)p(t)表示训练集中划分到该结点的样本的比例,例如100个样本,25个到了该结点,那p(t)=25100=14p \left( t \right)=\frac{25}{100}=\frac{1}{4}p(t)=10025=41,如果有缺失值,可以考虑用划分到该结点的样本的权值和除以全训练集的权重和
R(t)R\left( t \right)R(t)表示训练集中划分到该结点的样本的损失,可以用基尼指数、MSE、熵、错误率之类
Cα(t)=C(t)+α=p(t)R(t)+α C_{\alpha}\left( t \right) = C(t) + \alpha = p\left( t \right) R\left( t \right) + \alpha Cα(t)=C(t)+α=p(t)R(t)+α

当Cα(Tt)=Cα(t)C_{\alpha}\left( T_t \right) = C_{\alpha}\left( t \right)Cα(Tt)=Cα(t),说明他们发挥着差不多的作用,那一般子节点少的泛化性会高一点,就可以把他的子节点裁了

Cα(Tt)=Cα(t)⇒α=C(t)−C(Tt)∣Tt∣−1 C_{\alpha}\left( T_{t} \right) =C_{\alpha}\left( t \right) \Rightarrow \alpha = \frac{C\left( t \right) - C \left( T_{t} \right) }{\left| T_{t} \right| -1} Cα(Tt)=Cα(t)⇒α=∣Tt∣−1C(t)−C(Tt)

算法流程:

(1)设k=0,T=T0k=0, T=T_{0}k=0,T=T0

(2)设α=+inf⁡\alpha = +\infα=+inf

(3)自下而上地对各内部结点ttt计算C(Tt)C(T_{t})C(Tt),∣Tt∣\left| T_{t} \right|∣Tt∣以及
g(t)=C(t)−C(Tt)∣Tt∣−1 g\left( t \right) = \frac{C\left( t \right) - C \left( T_{t} \right) }{\left| T_{t} \right| -1} g(t)=∣Tt∣−1C(t)−C(Tt)
α=min⁡(α,g(t)) \alpha = \min \left( \alpha, g\left( t \right) \right) α=min(α,g(t))

(4)自上而下地访问内部结点ttt,如果有g(t)=αg\left( t \right) = \alphag(t)=α,进行剪枝,并对叶结点ttt以多数表决法决定其类,得到树TTT

(5)设k=k+1,αk=α,Tk=Tk = k +1, \alpha_{k} = \alpha, T_{k} =Tk=k+1,αk=α,Tk=T

(6)如果TTT不是由根节点单独构成的树,则回到步骤(4)

(7)采用交叉验证法在子树序列T0,T1,⋯ ,TnT_{0}, T_{1}, \cdots,T_{n}T0,T1,⋯,Tn中选取最优子树TαT_{\alpha}Tα

代码

id3

python 复制代码
#!/usr/bin/env python
# -*- encoding: utf-8 -*-
import numpy as np
import pandas as pd


class Node:
    def __init__(self, feature_name=None, category=None, majority_class=None):
        self.feature_name = feature_name # 特征
        self.category = category # 类别
        self.majority_class = majority_class # 多数类(备用类别)
        self.children = {} # 子节点,键为特征值,值为子节点

    def add_child(self, feature_value, child_node):
        self.children[feature_value] = child_node

    def is_leaf(self):
        return len(self.children) == 0
    
    def __repr__(self):
        if self.category:
            return f"Leaf({self.category})"
        return f"Node({self.feature_name}, majority={self.majority_class})"


def calc_entropy(df: pd.DataFrame, label_col: str) -> float:
    if df.empty:
        return 0.0

    counts = df[label_col].value_counts(normalize=True)
    # -(|C_k| / |D|) * log2(|C_k| / |D|)
    return -np.sum(counts * np.log2(counts))


def calc_feature_entropy(df: pd.DataFrame, feature_name: str, label_col: str) -> float:
    if df.empty:
        return 0.0
    counts = df[feature_name].value_counts(normalize=True)
    # (|D_i| / |D|) * entropy(D_i)
    return sum(
        p * calc_entropy(df[df[feature_name] == feature], label_col)
        for feature, p in counts.items()
    )


def calc_information_gain(
    total_entropy: float, df: pd.DataFrame, feature_name: str, label_col: str
) -> float:
    return total_entropy - calc_feature_entropy(df, feature_name, label_col)


def choose_best_feature(df: pd.DataFrame, label_col: str) -> str:
    total_entropy = calc_entropy(df, label_col)  # 预计算总熵
    best_gain = -1
    best_feature = None

    for feature in df.columns:
        if feature == label_col:
            continue
        gain = calc_information_gain(total_entropy, df, feature, label_col)
        print(f"Feature: {feature}, Information Gain: {gain}")
        if gain > best_gain:
            best_gain = gain
            best_feature = feature
    print(f"Best Feature: {best_feature}, Max Information Gain: {best_gain}")
    return best_feature, best_gain


class DecisionTreeID3:
    def __init__(self, epsilon: float = 1e-6):
        self.root = None
        self.epsilon = epsilon

    def _build(self, df: pd.DataFrame, label_col: str):
        # 全属于同一类
        labels = df[label_col].unique()
        if len(labels) == 1:
            return Node(category=labels[0])

        # 样本最多的类
        majority_class = df[label_col].mode()[0]
        # 没有特征可分
        if df.shape[1] == 1:
            return Node(majority_class=majority_class)

        best_feature, max_info_gain = choose_best_feature(df, label_col)
        node = Node(feature_name=best_feature, majority_class=majority_class)
        if max_info_gain < self.epsilon:
            return node

        for feature_value, subset in df.groupby(best_feature):
            child_node = self._build(subset.drop(columns=best_feature), label_col)
            node.add_child(feature_value, child_node)

        return node

    def fit(self, df: pd.DataFrame, label_col: str):
        if df.empty or label_col not in df.columns:
            raise ValueError("DataFrame is empty or label column is missing.")
        if df.isna().values.any():
            raise ValueError(
                "DataFrame contains NaN values. Please handle missing data before fitting the model."
            )
        numeric = df.select_dtypes(include=[np.number])
        if not np.isfinite(numeric.values).all():
            raise ValueError(
                "DataFrame contains non-finite values (inf, -inf). Please handle these values before fitting the model."
            )

        self.root = self._build(df, label_col)

    def predict_row(self, row: pd.Series) -> str:
        node = self.root
        while not node.is_leaf():
            if node.feature_name in row.index:
                feature_value = row[node.feature_name]
                if feature_value in node.children:
                    node = node.children[feature_value]
                else:
                    return node.majority_class
            else:
                print(
                    f"Warning: Feature {node.feature_name} not found in row, returning majority class."
                )
                return node.majority_class
        return node.category if node.category is not None else node.majority_class

    def predict(self, df: pd.DataFrame) -> pd.Series:
        if self.root is None:
            raise ValueError("The model has not been fitted yet.")
        return df.apply(self.predict_row, axis=1)
    
    def print_tree(self, node=None, indent="", feature_value=None):
        if node is None:
            if self.root is None:
                print("Tree not built yet.")
                return
            node = self.root
        
        # 打印当前节点
        prefix = f"{indent}{feature_value} → " if feature_value is not None else indent
        
        if node.is_leaf():
            value = node.category if node.category is not None else node.majority_class
            print(f"{prefix}{value}")
        else:
            print(f"{prefix}[{node.feature_name}] (majority: {node.majority_class})")
            # 递归打印子节点
            for value, child in node.children.items():
                self.print_tree(child, indent + "    ", value)


if __name__ == "__main__":
    df = pd.read_csv("PlayTennis.csv")
    df = pd.read_csv("watermelon.csv")
    df.drop(columns=["编号", "密度", "含糖率"], inplace=True)  # 删除不需要的列
    tree = DecisionTreeID3()
    tree.fit(df, label_col="好坏")
    tree.print_tree()
    predictions = tree.predict(df.drop(columns=["好坏"]))
    print("Predictions:")
    print(predictions)

C4.5

python 复制代码
#!/usr/bin/env python
# -*- encoding: utf-8 -*-
from collections import defaultdict
from queue import Queue

import numpy as np
import pandas as pd


class Node:
    def __init__(
        self, feature_name=None, threshold=None, category=None, majority_class=None
    ):
        self.feature_name = feature_name  # 特征
        self.threshold = threshold  # 阈值(用于连续特征)
        self.category = category  # 类别
        self.majority_class = majority_class  # 多数类(备用类别)
        self.ratio = {}  # 特征值对应的比例
        self.children = {}  # 子节点,键为特征值或者("<=", ">"),值为子节点

    def add_child(self, feature_value, child_node, child_ratio):
        self.children[feature_value] = child_node
        self.ratio[feature_value] = child_ratio

    def clear_child(self):
        self.ratio = {}
        self.children = {}

    def is_continuous(self):
        return self.threshold is not None

    def is_leaf(self):
        return len(self.children) == 0


def entropy(categories: pd.Series, weight: np.ndarray = None) -> float:
    if categories.empty:
        return 0.0

    if weight is None:
        counts = categories.value_counts(normalize=True)
        return -np.sum(counts * np.log2(counts))

    # 将类别转换为整数标签并计算加权频数
    labels, _ = pd.factorize(categories, sort=False)
    counts = np.bincount(labels, weights=weight, minlength=0)

    # 计算总权重和概率
    total_weight = counts.sum()
    if total_weight <= 0:
        return 0.0

    probabilities = counts / total_weight

    # 仅处理非零概率避免log(0)警告
    non_zero_probs = probabilities[probabilities > 0]
    return -np.sum(non_zero_probs * np.log2(non_zero_probs))


def choose_best_feature(df: pd.DataFrame, label_col: str, weight: np.ndarray) -> str:
    stats = []  # (特征, 信息增益, 信息增益率, 是否连续特征,阈值, 比例)
    if (weight <= 0).any():
        return None, None, None, None, None, None

    weight_sum = weight.sum()

    for feature in df.columns:
        if feature == label_col:
            continue
        feature_mask = df[feature].notna()
        weight_feature = weight[feature_mask]  # 过滤掉缺失值的权重
        df_feature = df.loc[feature_mask]  # \tilde{D}
        weight_feature_sum = weight_feature.sum()
        rho = weight_feature_sum / weight_sum  # 权重
        total_entropy = entropy(
            df_feature[label_col], weight_feature
        )  # entropy(\tilde{D})
        # 连续特征
        if pd.api.types.is_numeric_dtype(df_feature[feature]):
            values = df_feature[feature].sort_values().unique()
            if len(values) <= 1:  # 值不足无法划分
                continue
            thresholds = (values[:-1] + values[1:]) / 2  # 划分点

            max_info_gain = -np.inf  # 最大信息增益
            best_info_gain_rate = -np.inf  # 最佳信息增益率
            best_threshold = None  # 最佳划分点
            best_rk = None

            for threshold in thresholds:
                left_mask = df_feature[feature] <= threshold
                left_weight = weight_feature[left_mask]
                left_weight_sum = left_weight.sum()

                right_mask = ~left_mask
                right_weight = weight_feature[right_mask]
                right_weight_sum = right_weight.sum()

                rk_left = left_weight_sum / weight_feature_sum
                rk_right = right_weight_sum / weight_feature_sum

                left_entropy = entropy(
                    df_feature.loc[left_mask, label_col], left_weight
                )
                right_entropy = entropy(
                    df_feature.loc[right_mask, label_col], right_weight
                )

                feature_entropy = rk_left * left_entropy + rk_right * right_entropy

                info_gain = rho * (total_entropy - feature_entropy)
                if info_gain > max_info_gain:
                    max_info_gain = info_gain
                    best_threshold = threshold
                    iv = rho * (
                        -rk_left * np.log2(rk_left) - rk_right * np.log2(rk_right)
                    )  # 固有值
                    if iv <= 0:
                        continue
                    best_info_gain_rate = info_gain / iv
                    best_rk = {"<=": rk_left, ">": rk_right}
            if max_info_gain > -np.inf:  # 仅添加有效划分点
                stats.append(
                    (
                        feature,
                        max_info_gain,
                        best_info_gain_rate,
                        True,
                        best_threshold,
                        best_rk,
                    )
                )
        else:
            # 离散特征
            feature_values = df_feature[feature].unique()
            if len(feature_values) <= 1:
                continue
            all_rk = {}
            # r_k * entropy(\tilde{D}^v)
            feature_entropy = 0.0
            for value in feature_values:
                mask = df_feature[feature] == value
                value_weight = weight_feature[mask]
                value_entropy = entropy(df_feature.loc[mask, label_col], value_weight)
                rk = value_weight.sum() / weight_feature_sum
                all_rk[value] = rk
                feature_entropy += rk * value_entropy
            info_gain = rho * (total_entropy - feature_entropy)
            iv = rho * entropy(df_feature[feature], weight_feature)  # 固有值
            if iv <= 0:
                continue
            info_gain_rate = info_gain / iv
            stats.append((feature, info_gain, info_gain_rate, False, None, all_rk))

    # 按照信息增益率排序,选择最优特征
    if not stats:
        return None, None, None, None, None, None
    mean_info_gain = np.mean([s[1] for s in stats])
    useful_stats = [s for s in stats if s[1] > mean_info_gain]
    # if not useful_stats:
    #     return max(stats, key=lambda x: x[2])
    return max(useful_stats or stats, key=lambda x: x[2])


class DecisionTreeC45:
    def __init__(
        self,
        epsilon: float = 1e-6,
        pre_prune=False,
        post_prune=False,
        reuse_feature: bool = False,
    ):
        self.root = None
        self.epsilon = epsilon
        self.pre_prune = pre_prune
        self.post_prune = post_prune
        self.reuse_feature = reuse_feature

    def _build(
        self,
        df: pd.DataFrame,
        label_col: str,
        weight: np.ndarray,
        root=None,
        val_df=None,
    ) -> Node:
        # 全属于同一类
        labels = df[label_col].unique()
        if len(labels) == 1:
            return Node(category=labels[0])

        # 样本最多的类
        majority_class = df[label_col].mode()[0]
        # 没有特征可分
        if df.shape[1] == 1:
            return Node(majority_class=majority_class)

        (
            best_feature,
            max_info_gain,
            best_info_gain_rate,
            is_continuous,
            threshold,
            best_rk,
        ) = choose_best_feature(df, label_col, weight)
        node = Node(
            feature_name=best_feature,
            threshold=threshold,
            majority_class=majority_class,
        )
        if root is None:
            root = node

        if best_feature is None or best_info_gain_rate < self.epsilon:
            # return Node(majority_class=majority_class)
            return node

        # 预剪枝
        if val_df and self.pre_prune:
            pre_acc = self.calculate_acc(val_df, label_col, root)
            # 创建临时子树
            if is_continuous:
                node.add_child(
                    "<=",
                    Node(
                        majority_class=df.loc[
                            df[best_feature] <= threshold, label_col
                        ].mode()[0]
                    ),
                    best_rk["<="],
                )
                node.add_child(
                    ">",
                    Node(
                        majority_class=df.loc[
                            df[best_feature] > threshold, label_col
                        ].mode()[0]
                    ),
                    best_rk[">"],
                )
            else:
                for value, rk in best_rk.items():
                    node.add_child(
                        value,
                        Node(
                            majority_class=df.loc[
                                df[best_feature] == value, label_col
                            ].mode()[0]
                        ),
                        rk,
                    )
            prune_acc = self.calculate_acc(val_df, label_col, root)
            # 清除临时子树
            node.clear_child()
            # 如果剪枝后准确率没有提升,则终止划分
            if pre_acc >= prune_acc:
                return node

        na_mask = df[best_feature].isna()

        if is_continuous:  # 连续特征
            left_mask = df[best_feature] <= threshold  # 小于等于阈值的样本(不包含na)
            right_mask = df[best_feature] > threshold

            left_weight = weight.copy()
            left_weight[na_mask] *= best_rk["<="]  # NA权重分配给左子树
            left_weight = left_weight[left_mask | na_mask]

            right_weight = weight.copy()
            right_weight[na_mask] *= best_rk[">"]
            right_weight = right_weight[right_mask | na_mask]

            if not self.reuse_feature:
                df = df.drop(columns=[best_feature])

            left_df = df.loc[left_mask | na_mask]
            if self.reuse_feature and df.loc[left_mask, best_feature].nunique() == 1:
                left_df = left_df.drop(columns=[best_feature])

            right_df = df.loc[right_mask | na_mask]
            if self.reuse_feature and df.loc[right_mask, best_feature].nunique() == 1:
                right_df = right_df.drop(columns=[best_feature])

            left_node = self._build(left_df, label_col, left_weight, root, val_df)
            right_node = self._build(right_df, label_col, right_weight, root, val_df)
            node.add_child("<=", left_node, best_rk["<="])
            node.add_child(">", right_node, best_rk[">"])
        else:  # 离散特征
            for value, rk in best_rk.items():
                mask = df[best_feature] == value

                cur_weight = weight.copy()
                cur_weight[na_mask] *= rk
                cur_weight = cur_weight[mask | na_mask]

                cur_df = df.loc[mask | na_mask].drop(columns=[best_feature])
                next_node = self._build(cur_df, label_col, cur_weight, root, val_df)
                node.add_child(value, next_node, rk)

        # 后剪枝
        if val_df and self.post_prune and node.children:
            flag = True
            for child in node.children.values():
                if not child.is_leaf():
                    flag = False
                    break
            if not flag:
                return node
            pre_acc = self.calculate_acc(val_df, label_col, root)
            children = node.children
            ratio = node.ratio
            node.clear_child()
            prune_acc = self.calculate_acc(val_df, label_col, root)
            if pre_acc >= prune_acc:
                return node
            node.children = children
            node.ratio = ratio

        return node

    def fit(self, df: pd.DataFrame, label_col: str, val_df=None, weight=None):
        if df.empty or label_col not in df.columns:
            raise ValueError("DataFrame is empty or label column is missing.")

        if df[label_col].isna().any():
            raise ValueError("Label column contains NaN values.")

        if weight is None:
            weight = np.ones(len(df))  # 初始权重为1
        elif (weight <= 0).any():
            raise ValueError("Weight is not positive.")

        weight = np.ones(len(df))  # 初始权重为1
        self.root = self._build(df, label_col, weight, None, val_df)

    def predict_row(self, row: pd.Series, node) -> str:
        q = Queue()
        q.put((node, 1.0))
        res = defaultdict(float)
        while not q.empty():
            current_node, current_weight = q.get()
            if current_node.is_leaf():
                res[
                    current_node.category
                    if current_node.category is not None
                    else current_node.majority_class
                ] += current_weight
                continue

            if current_node.feature_name not in row.index:
                print(
                    f"Warning: Feature {current_node.feature_name} not found in row, returning majority class."
                )
                res[current_node.majority_class] += current_weight
                continue

            feature_value = row[current_node.feature_name]
            if pd.notna(feature_value):
                if current_node.is_continuous():
                    if feature_value <= current_node.threshold:
                        child_node = current_node.children["<="]
                    else:
                        child_node = current_node.children[">"]
                    q.put((child_node, current_weight))
                else:
                    if feature_value in current_node.children:
                        child_node = current_node.children[feature_value]
                        q.put((child_node, current_weight))
                    else:
                        # 如果特征值不在子节点中,返回多数类
                        print(
                            f"Warning: Feature {current_node.feature_name} has value {feature_value} not found in children, returning majority class."
                        )
                        res[current_node.majority_class] += current_weight
            else:
                for value, child_node in current_node.children.items():
                    q.put((child_node, current_weight * current_node.ratio[value]))

        # Combine results from all paths
        if not res:
            return self.root.majority_class
        return max(res, key=res.get)

    def calculate_acc(self, df, label_col, node):
        if not df or df.empty:
            return 0.0
        preds = df.apply(lambda row: self.predict_row(row, node), axis=1)
        return (preds == df[label_col]).mean()

    def predict(self, df: pd.DataFrame) -> pd.Series:
        if self.root is None:
            raise ValueError("The model has not been fitted yet.")
        return df.apply(lambda row: self.predict_row(row, self.root), axis=1)

    def print_tree(self, node=None, indent: str = "", prefix: str = ""):
        """打印决策树结构"""
        if node is None:
            node = self.root
            print("Decision Tree:")

        # 叶子节点打印类别
        if node.is_leaf():
            leaf_class = node.category if node.category else node.majority_class
            print(f"{indent}{prefix} Leaf: {leaf_class}")
            return

        # 内部节点打印特征
        if node.is_continuous():
            print(f"{indent}{prefix} {node.feature_name} <= {node.threshold:.3f}")
            # 递归打印子节点
            self.print_tree(node.children.get("<="), indent + "  ", "├── <=: ")
            self.print_tree(node.children.get(">"), indent + "  ", "└── >: ")
        else:
            print(f"{indent}{prefix} {node.feature_name}")
            # 递归打印所有子节点
            children = list(node.children.items())
            for i, (value, child_node) in enumerate(children):
                last_child = i == len(children) - 1
                branch = "└──" if last_child else "├──"
                self.print_tree(child_node, indent + "  ", f"{branch} {value}: ")


if __name__ == "__main__":
    df = pd.read_csv("watermelon_na.csv")
    df.drop(columns=["编号"], inplace=True)  # 删除不需要的列

    tree = DecisionTreeC45(epsilon=1e-6, reuse_feature=False)
    tree.fit(df, label_col="好坏")
    tree.print_tree()

Cart

python 复制代码
#!/usr/bin/env python
# -*- encoding: utf-8 -*-
from collections import defaultdict
from copy import deepcopy
from queue import Queue

import numpy as np
import pandas as pd


class Node:
    def __init__(
        self,
        feature_name=None,
        threshold=None,
        category=None,
        majority_class=None,
        is_continuous=False,
        node_weight=None,
        gini=None,
        R=None,
        R_T=None,
        leaf_count=None,
    ):
        self.feature_name = feature_name  # 特征
        self.threshold = threshold  # 阈值
        self.category = category  # 类别
        self.majority_class = majority_class  # 多数类(备用类别)
        self.is_continuous = is_continuous  # 连续
        self.ratio = {}  # 特征值对应的比例
        self.children = {}  # 子节点,键为特征值或者("<=", ">"),值为子节点 (>表示大于或者!=)
        self.node_weight = node_weight  # 节点权重(样本权重和)
        self.gini = gini  # 节点基尼指数
        self.R = R  # 节点的不纯度代价 R(t)
        self.R_T = R_T  # 子树的不纯度代价 R(T_t)
        self.leaf_count = leaf_count  # 子树叶子节点数量

    def add_child(self, feature_value, child_node, child_ratio):
        self.children[feature_value] = child_node
        self.ratio[feature_value] = child_ratio

    def clear_child(self):
        self.ratio = {}
        self.children = {}
        # 剪枝后变为叶节点,R_T等于节点自身R
        self.R_T = self.R_T
        self.leaf_count = 1

    def is_leaf(self):
        return len(self.children) == 0


def gini(categories: pd.Series, weight: np.ndarray = None) -> float:
    if categories.empty:
        return 0.0
    if weight is None:
        counts = categories.value_counts(normalize=True)
        # 1 - sum(p_k^2)
        return 1 - np.sum(counts**2)

    # 将类别转换为整数标签并计算加权频数
    labels, _ = pd.factorize(categories, sort=False)
    counts = np.bincount(labels, weights=weight, minlength=0)

    # 计算总权重和概率
    total_weight = counts.sum()
    if total_weight <= 0:
        return 0.0

    probabilities = counts / total_weight
    return 1 - np.sum(probabilities**2)


def choose_best_feature(df: pd.DataFrame, label_col: str, weight: np.ndarray) -> str:
    if (weight <= 0).any():
        return None, None, None, None, None

    best_feature = None
    min_gini_index = np.inf
    best_is_continuous = None
    best_threshold = None
    best_rk = None

    weight_sum = weight.sum()
    for feature in df.columns:
        if feature == label_col:
            continue
        feature_mask = df[feature].notna()
        weight_feature = weight[feature_mask]  # 过滤掉缺失值的权重
        df_feature = df.loc[feature_mask]  # \tilde{D}
        weight_feature_sum = weight_feature.sum()
        rho = weight_feature_sum / weight_sum  # 权重

        is_continuous = pd.api.types.is_numeric_dtype(df_feature[feature])
        splits = None
        if is_continuous:
            values = df_feature[feature].sort_values().unique()
            if len(values) <= 1:  # 值不足无法划分
                continue
            splits = (values[:-1] + values[1:]) / 2  # 划分点
        else:
            feature_values = df_feature[feature].unique()
            if len(feature_values) <= 1:
                continue
            splits = feature_values

        for value in splits:
            left_mask = (
                (df_feature[feature] <= value)
                if is_continuous
                else (df_feature[feature] == value)
            )
            left_weight = weight_feature[left_mask]
            left_weight_sum = left_weight.sum()

            right_mask = ~left_mask
            right_weight = weight_feature[right_mask]
            right_weight_sum = right_weight.sum()

            rk_left = left_weight_sum / weight_feature_sum
            rk_right = right_weight_sum / weight_feature_sum

            left_gini = gini(df_feature.loc[left_mask, label_col], left_weight)
            right_gini = gini(df_feature.loc[right_mask, label_col], right_weight)
            feature_gini = rk_left * left_gini + rk_right * right_gini
            # rho * (rk * gini(\tilde{D}^{v}))
            gini_index = rho * feature_gini

            if gini_index < min_gini_index:
                best_feature = feature
                min_gini_index = gini_index
                best_is_continuous = is_continuous
                best_threshold = value
                best_rk = {"<=": rk_left, ">": rk_right}
    return best_feature, min_gini_index, best_is_continuous, best_threshold, best_rk


class DecisionTreeCart:
    def __init__(
        self,
        epsilon: float = 1e-6,
        pre_prune=False,
        post_prune=False,
        reuse_feature: bool = True,
        ccp_alpha=0.0,
    ):
        self.root = None
        self.epsilon = epsilon
        self.pre_prune = pre_prune
        self.post_prune = post_prune
        self.reuse_feature = reuse_feature
        self.ccp_alpha = ccp_alpha

    def _build(
        self,
        df: pd.DataFrame,
        label_col: str,
        weight: np.ndarray,
        root=None,
        val_df=None,
        total_weight=None,
    ) -> Node:
        # 计算当前节点权重和基尼指数
        node_weight = weight.sum() / total_weight
        node_gini = gini(df[label_col], weight)
        node_R = node_weight * node_gini
        # 全属于同一类
        labels = df[label_col].unique()
        if len(labels) == 1:
            return Node(
                category=labels[0],
                node_weight=node_weight,
                gini=node_gini,
                R=node_R,
                R_T=node_R,
                leaf_count=1,
            )

        # 样本最多的类
        majority_class = df[label_col].mode()[0]
        # 没有特征可分
        if df.shape[1] == 1:
            return Node(
                majority_class=majority_class,
                node_weight=node_weight,
                gini=node_gini,
                R=node_R,
                R_T=node_R,
                leaf_count=1,
            )

        best_feature, min_gini_index, best_is_continuous, best_threshold, best_rk = (
            choose_best_feature(df, label_col, weight)
        )

        node = Node(
            feature_name=best_feature,
            threshold=best_threshold,
            majority_class=majority_class,
            is_continuous=best_is_continuous,
            node_weight=node_weight,
            gini=node_gini,
            R=node_R,
        )

        if root is None:
            root = node

        if best_feature is None or min_gini_index < self.epsilon:
            # return Node(majority_class=majority_class)
            node.R_T = node_R
            node.leaf_count = 1
            return node

        na_mask = df[best_feature].isna()
        left_mask = None
        if best_is_continuous:
            left_mask = (
                df[best_feature] <= best_threshold
            )  # 小于等于阈值的样本(不包含na)
        else:
            left_mask = df[best_feature] == best_threshold  # 等于阈值的样本(不包含na)
        right_mask = (~left_mask) & (~na_mask)  # 其他样本(不包含na)

        # 预剪枝
        if val_df and self.pre_prune:
            pre_acc = self.calculate_acc(val_df, label_col, root)
            # 创建临时子树
            node.add_child(
                "<=",
                Node(majority_class=df.loc[left_mask, label_col].mode()[0]),
                best_rk["<="],
            )
            node.add_child(
                ">",
                Node(majority_class=df.loc[right_mask, label_col].mode()[0]),
                best_rk[">"],
            )
            prune_acc = self.calculate_acc(val_df, label_col, root)
            # 清除临时子树
            node.clear_child()
            # 如果剪枝后准确率没有提升,则终止划分
            if pre_acc >= prune_acc:
                node.R_T = node_R
                node.leaf_count = 1
                return node

        left_weight = weight.copy()
        left_weight[na_mask] *= best_rk["<="]  # NA权重分配给左子树
        left_weight = left_weight[left_mask | na_mask]

        right_weight = weight.copy()
        right_weight[na_mask] *= best_rk[">"]
        right_weight = right_weight[right_mask | na_mask]

        if not self.reuse_feature:
            df = df.drop(columns=[best_feature])
        left_df = df.loc[left_mask | na_mask]
        if self.reuse_feature and df.loc[left_mask, best_feature].nunique() == 1:
            left_df = left_df.drop(columns=[best_feature])

        right_df = df.loc[right_mask | na_mask]
        if self.reuse_feature and df.loc[right_mask, best_feature].nunique() == 1:
            right_df = right_df.drop(columns=[best_feature])

        left_node = self._build(
            left_df, label_col, left_weight, root, val_df, total_weight
        )
        right_node = self._build(
            right_df, label_col, right_weight, root, val_df, total_weight
        )
        node.add_child("<=", left_node, best_rk["<="])
        node.add_child(">", right_node, best_rk[">"])

        node.R_T = left_node.R_T + right_node.R_T
        node.leaf_count = left_node.leaf_count + right_node.leaf_count
        return node

    def fit(self, df: pd.DataFrame, label_col: str, val_df=None, weight=None):
        if df.empty or label_col not in df.columns:
            raise ValueError("DataFrame is empty or label column is missing.")

        if df[label_col].isna().any():
            raise ValueError("Label column contains NaN values.")

        if weight is None:
            weight = np.ones(len(df))  # 初始权重为1
        elif (weight <= 0).any():
            raise ValueError("Weight is not positive.")
        total_weight = weight.sum()
        self.root = self._build(df, label_col, weight, None, val_df, total_weight)

        if self.post_prune and self.ccp_alpha > 0 and self.val_df:
            self._cost_complexity_pruning(val_df, label_col)

    def _cost_complexity_pruning(self, val_df, label_col):
        """
        执行代价复杂度剪枝算法
        步骤:
        1. 初始化: k=0, T=T_0, alpha=+inf
        2. 自底向上计算每个内部节点t的g(t)
        3. 找到最小的g(t)作为alpha_k,并剪枝对应的节点
        4. 重复直到只剩根节点
        5. 使用验证集选择最优子树
        """
        # 步骤1: 初始化
        k = 0
        T = deepcopy(self.root)  # 当前子树
        alpha = float("inf")
        self.pruned_trees = []  # 存储子树序列 T0, T1, ..., Tn
        self.alphas = []  # 存储对应的alpha序列

        # 添加原始树
        self.pruned_trees.append(deepcopy(T))
        self.alphas.append(alpha)

        # 步骤2-6: 循环剪枝直到只剩根节点
        while not T.is_leaf():
            min_g = float("inf")
            min_g_nodes = []

            q = Queue()
            q.put(T)

            while not q.empty():
                p = q.get()
                for child in p.children.values():
                    if child.is_leaf():
                        continue
                    assert child.leaf_count > 1
                    q.put(child)
                    cur_g = (child.R - child.R_T) / (child.leaf_count - 1)
                    if cur_g < min_g:
                        min_g = cur_g
                        min_g_nodes.append(child)
                    elif abs(cur_g - min_g) < 1e-6:  # cur_g == min_g
                        min_g_nodes.append(child)

            # 如果没有找到可剪枝的节点,停止循环
            if not min_g_nodes:
                break

            # 更新alpha
            alpha = min_g

            # 剪枝所有g(t)=alpha的节点
            for node in min_g_nodes:
                node.clear_child()

            # 更新整棵树的R_T和leaf_count
            self._update_tree_metrics(T)

            # 保存当前子树和alpha
            k += 1
            self.pruned_trees.append(deepcopy(T))
            self.alphas.append(alpha)

            # 如果达到ccp_alpha阈值则停止
            if alpha > self.ccp_alpha:
                break

        # 步骤7: 使用验证集选择最优子树
        if val_df is not None:
            best_acc = -1
            best_tree_idx = -1

            for i, tree in enumerate(self.pruned_trees):
                acc = self.calculate_acc(val_df, label_col, tree)
                if acc > best_acc:
                    best_acc = acc
                    best_tree_idx = i

            # 选择最优子树
            self.root = self.pruned_trees[best_tree_idx]
            print(
                f"Selected subtree T_{best_tree_idx} with alpha={self.alphas[best_tree_idx]:.6f}, accuracy={best_acc:.4f}"
            )

    def _update_tree_metrics(self, node):
        if node.is_leaf():
            return

        total_R_T = 0.0
        total_leaf_count = 0
        for child in node.children.values():
            self._update_tree_metrics(child)
            total_R_T += child.R_T
            total_leaf_count += child.leaf_count

        node.R_T = total_R_T
        node.leaf_count = total_leaf_count

    def predict_row(self, row: pd.Series, node) -> str:
        q = Queue()
        q.put((node, 1.0))
        res = defaultdict(float)
        while not q.empty():
            current_node, current_weight = q.get()
            if current_node.is_leaf():
                res[
                    current_node.category
                    if current_node.category is not None
                    else current_node.majority_class
                ] += current_weight
                continue

            if current_node.feature_name not in row.index:
                print(
                    f"Warning: Feature {current_node.feature_name} not found in row, returning majority class."
                )
                res[current_node.majority_class] += current_weight
                continue

            feature_value = row[current_node.feature_name]
            if pd.notna(feature_value):
                if current_node.is_continuous:
                    if feature_value <= current_node.threshold:
                        child_node = current_node.children["<="]
                    else:
                        child_node = current_node.children[">"]
                    q.put((child_node, current_weight))
                else:
                    if feature_value == current_node.threshold:
                        child_node = current_node.children["<="]

                    else:
                        child_node = current_node.children[">"]
                    q.put((child_node, current_weight))
            else:
                for value, child_node in current_node.children.items():
                    q.put((child_node, current_weight * current_node.ratio[value]))

        # Combine results from all paths
        if not res:
            return self.root.majority_class
        return max(res, key=res.get)

    def calculate_acc(self, df, label_col, node):
        if not df or df.empty:
            return 0.0
        preds = df.apply(lambda row: self.predict_row(row, node), axis=1)
        return (preds == df[label_col]).mean()

    def predict(self, df: pd.DataFrame) -> pd.Series:
        if self.root is None:
            raise ValueError("The model has not been fitted yet.")
        return df.apply(lambda row: self.predict_row(row, self.root), axis=1)

    def print_tree(self, node=None, indent: str = "", prefix: str = ""):
        """打印决策树结构"""
        if node is None:
            node = self.root
            print("Decision Tree:")
            print(f"Total leaves: {node.leaf_count}")

        # 叶子节点打印类别
        if node.is_leaf():
            leaf_class = node.category if node.category else node.majority_class
            print(f"{indent}{prefix} Leaf: {leaf_class} "
                  f"(R={node.R:.4f}, samples={node.node_weight:.3f})")
            return

        # 内部节点打印特征
        if node.is_continuous:
            print(f"{indent}{prefix} {node.feature_name} <= {node.threshold:.3f} "
                  f"[R_T={node.R_T:.4f}, leaves={node.leaf_count}]")
            # 递归打印子节点
            self.print_tree(node.children.get("<="), indent + "  ", "├── <=: ")
            self.print_tree(node.children.get(">"), indent + "  ", "└── >: ")
        else:
            print(f"{indent}{prefix} {node.feature_name} == {node.threshold} "
                  f"[R_T={node.R_T:.4f}, leaves={node.leaf_count}]")
            self.print_tree(node.children.get("<="), indent + "  ", "├── ==: ")
            self.print_tree(node.children.get(">"), indent + "  ", "└── !=: ")


if __name__ == "__main__":
    df = pd.read_csv("watermelon.csv")
    df.drop(columns=["编号"], inplace=True)  # 删除不需要的列

    tree = DecisionTreeCart(
        epsilon=1e-6, pre_prune=False, post_prune=False, reuse_feature=True, ccp_alpha=0
    )
    tree.fit(df, label_col="好坏")
    tree.print_tree()

Cart回归

python 复制代码
#!/usr/bin/env python
# -*- encoding: utf-8 -*-
from copy import deepcopy
from queue import Queue

import numpy as np
import pandas as pd


class Node:
    def __init__(
        self,
        feature_name=None,
        threshold=None,
        value=None,
        is_continuous=False,
        node_weight=None,
        gini=None,
        R=None,
        R_T=None,
        leaf_count=None,
    ):
        self.feature_name = feature_name  # 特征
        self.threshold = threshold  # 阈值
        self.value = value  # 叶节点的预测值
        self.is_continuous = is_continuous  # 连续
        self.ratio = {}  # 特征值对应的比例
        self.children = {}  # 子节点,键为特征值或者("<=", ">"),值为子节点 (>表示大于或者!=)
        self.node_weight = node_weight  # 节点权重(样本权重和)
        self.gini = gini  # 节点基尼指数
        self.R = R  # 节点的不纯度代价 R(t)
        self.R_T = R_T  # 子树的不纯度代价 R(T_t)
        self.leaf_count = leaf_count  # 子树叶子节点数量

    def add_child(self, feature_value, child_node, child_ratio):
        self.children[feature_value] = child_node
        self.ratio[feature_value] = child_ratio

    def clear_child(self):
        self.ratio = {}
        self.children = {}
        # 剪枝后变为叶节点,R_T等于节点自身R
        self.R_T = self.R_T
        self.leaf_count = 1

    def is_leaf(self):
        return len(self.children) == 0


def calculate_r(y: np.ndarray, weight: np.ndarray = None):
    if len(y) == 0:
        return 0.0

    if weight is None:
        weight = np.ones(len(y))

    if weight.sum() <= 0:
        return 0.0

    # min_c \sum w_i (y_i - c)^2 = (y - c1)^T W (y - c1) => c = (1^T W y) / (1^T W 1) = (\sum w_i y_i) / (\sum w_i)
    c = np.average(y, weights=weight)
    # \sum w_i (y_i - c)^2
    return np.sum(weight * ((y - c) ** 2))


def choose_best_feature(df: pd.DataFrame, label_col: str, weight: np.ndarray) -> str:
    if (weight <= 0).any():
        return None, None, None, None, None

    best_feature = None
    min_mse = np.inf
    best_is_continuous = None
    best_threshold = None
    best_rk = None

    weight_sum = weight.sum()
    for feature in df.columns:
        if feature == label_col:
            continue
        feature_mask = df[feature].notna()
        weight_feature = weight[feature_mask]  # 过滤掉缺失值的权重
        df_feature = df.loc[feature_mask]  # \tilde{D}
        weight_feature_sum = weight_feature.sum()
        rho = weight_feature_sum / weight_sum  # 权重

        is_continuous = pd.api.types.is_numeric_dtype(df_feature[feature])
        splits = None
        if is_continuous:
            values = df_feature[feature].sort_values().unique()
            if len(values) <= 1:  # 值不足无法划分
                continue
            splits = (values[:-1] + values[1:]) / 2  # 划分点
        else:
            feature_values = df_feature[feature].unique()
            if len(feature_values) <= 1:
                continue
            splits = feature_values

        for value in splits:
            left_mask = (
                (df_feature[feature] <= value)
                if is_continuous
                else (df_feature[feature] == value)
            )
            left_weight = weight_feature[left_mask]
            left_weight_sum = left_weight.sum()

            right_mask = ~left_mask
            right_weight = weight_feature[right_mask]
            right_weight_sum = right_weight.sum()

            rk_left = left_weight_sum / weight_feature_sum
            rk_right = right_weight_sum / weight_feature_sum

            left_mse = calculate_r(
                df_feature.loc[left_mask, label_col].to_numpy(), left_weight
            )
            right_mse = calculate_r(
                df_feature.loc[right_mask, label_col].to_numpy(), right_weight
            )
            feature_mse = rk_left * left_mse + rk_right * right_mse
            # rho * (rk * gini(\tilde{D}^{v}))
            weighted_mse = rho * feature_mse

            if weighted_mse < min_mse:
                best_feature = feature
                min_mse = weighted_mse
                best_is_continuous = is_continuous
                best_threshold = value
                best_rk = {"<=": rk_left, ">": rk_right}
    return best_feature, min_mse, best_is_continuous, best_threshold, best_rk


class DecisionTreeCartRegression:
    def __init__(
        self,
        epsilon: float = 1e-6,
        pre_prune=False,
        post_prune=False,
        reuse_feature: bool = True,
        ccp_alpha=0.0,
    ):
        self.root = None
        self.epsilon = epsilon
        self.pre_prune = pre_prune
        self.post_prune = post_prune
        self.reuse_feature = reuse_feature
        self.ccp_alpha = ccp_alpha

    def _build(
        self,
        df: pd.DataFrame,
        label_col: str,
        weight: np.ndarray,
        root=None,
        val_df=None,
        total_weight=None,
    ) -> Node:
        # 计算当前节点权重和基尼指数
        node_weight = weight.sum() / total_weight
        node_value = np.average(df[label_col].to_numpy(), weights=weight)
        node_R = node_weight * calculate_r(df[label_col].to_numpy(), weight)

        # y一样或者没有特征可分
        if df[label_col].max() - df[label_col].min() < 1e-6 or df.shape[1] == 1:
            return Node(
                value=node_value,
                node_weight=node_weight,
                gini=node_value,
                R=node_R,
                R_T=node_R,
                leaf_count=1,
            )

        best_feature, min_gini_index, best_is_continuous, best_threshold, best_rk = (
            choose_best_feature(df, label_col, weight)
        )

        node = Node(
            feature_name=best_feature,
            threshold=best_threshold,
            value=node_value,
            is_continuous=best_is_continuous,
            node_weight=node_weight,
            gini=node_value,
            R=node_R,
        )

        if root is None:
            root = node

        if best_feature is None:
            # return Node(majority_class=majority_class)
            node.R_T = node_R
            node.leaf_count = 1
            return node

        na_mask = df[best_feature].isna()
        left_mask = None
        if best_is_continuous:
            left_mask = (
                df[best_feature] <= best_threshold
            )  # 小于等于阈值的样本(不包含na)
        else:
            left_mask = df[best_feature] == best_threshold  # 等于阈值的样本(不包含na)
        right_mask = (~left_mask) & (~na_mask)  # 其他样本(不包含na)

        # 预剪枝
        if val_df and self.pre_prune:
            pre_mse = self.calculate_mse_from_node(val_df, label_col, root)
            # 创建临时子树(一层)
            left_value = (
                np.average(
                    df.loc[left_mask, label_col].to_numpy(), weights=weight[left_mask]
                )
                if left_mask.any()
                else node_value
            )

            right_value = (
                np.average(
                    df.loc[right_mask, label_col].to_numpy(), weights=weight[right_mask]
                )
                if right_mask.any()
                else node_value
            )

            node.add_child("<=", Node(value=left_value), best_rk["<="])
            node.add_child(">", Node(value=right_value), best_rk[">"])
            prune_mse = self.calculate_mse_from_node(val_df, label_col, root)
            # 清除临时子树
            node.clear_child()
            # 如果剪枝后准确率没有提升,则终止划分
            if pre_mse <= prune_mse:
                node.R_T = node_R
                node.leaf_count = 1
                return node

        left_weight = weight.copy()
        left_weight[na_mask] *= best_rk["<="]  # NA权重分配给左子树
        left_weight = left_weight[left_mask | na_mask]

        right_weight = weight.copy()
        right_weight[na_mask] *= best_rk[">"]
        right_weight = right_weight[right_mask | na_mask]

        if not self.reuse_feature:
            df = df.drop(columns=[best_feature])
        left_df = df.loc[left_mask | na_mask]
        if self.reuse_feature and df.loc[left_mask, best_feature].nunique() == 1:
            left_df = left_df.drop(columns=[best_feature])

        right_df = df.loc[right_mask | na_mask]
        if self.reuse_feature and df.loc[right_mask, best_feature].nunique() == 1:
            right_df = right_df.drop(columns=[best_feature])

        left_node = self._build(
            left_df, label_col, left_weight, root, val_df, total_weight
        )
        right_node = self._build(
            right_df, label_col, right_weight, root, val_df, total_weight
        )
        node.add_child("<=", left_node, best_rk["<="])
        node.add_child(">", right_node, best_rk[">"])

        node.R_T = left_node.R_T + right_node.R_T
        node.leaf_count = left_node.leaf_count + right_node.leaf_count
        return node

    def fit(self, df: pd.DataFrame, label_col: str, val_df=None, weight=None):
        if df.empty or label_col not in df.columns:
            raise ValueError("DataFrame is empty or label column is missing.")

        if df[label_col].isna().any():
            raise ValueError("y column contains NaN values.")

        if not pd.api.types.is_numeric_dtype(df[label_col]):
            raise ValueError("y column only support numeric values.")

        if weight is None:
            weight = np.ones(len(df))  # 初始权重为1
        elif (weight <= 0).any():
            raise ValueError("Weight is not positive.")
        total_weight = weight.sum()
        self.root = self._build(df, label_col, weight, None, val_df, total_weight)

        if self.post_prune and self.ccp_alpha > 0 and self.val_df:
            self._cost_complexity_pruning(val_df, label_col)

    def _cost_complexity_pruning(self, val_df, label_col):
        """
        执行代价复杂度剪枝算法
        步骤:
        1. 初始化: k=0, T=T_0, alpha=+inf
        2. 自底向上计算每个内部节点t的g(t)
        3. 找到最小的g(t)作为alpha_k,并剪枝对应的节点
        4. 重复直到只剩根节点
        5. 使用验证集选择最优子树
        """
        # 步骤1: 初始化
        k = 0
        T = deepcopy(self.root)  # 当前子树
        alpha = float("inf")
        self.pruned_trees = []  # 存储子树序列 T0, T1, ..., Tn
        self.alphas = []  # 存储对应的alpha序列

        # 添加原始树
        self.pruned_trees.append(deepcopy(T))
        self.alphas.append(alpha)

        # 步骤2-6: 循环剪枝直到只剩根节点
        while not T.is_leaf():
            min_g = float("inf")
            min_g_nodes = []

            q = Queue()
            q.put(T)

            while not q.empty():
                p = q.get()
                for child in p.children.values():
                    if child.is_leaf():
                        continue
                    assert child.leaf_count > 1
                    q.put(child)
                    cur_g = (child.R - child.R_T) / (child.leaf_count - 1)
                    if cur_g < min_g:
                        min_g = cur_g
                        min_g_nodes.append(child)
                    elif abs(cur_g - min_g) < 1e-6:  # cur_g == min_g
                        min_g_nodes.append(child)

            # 如果没有找到可剪枝的节点,停止循环
            if not min_g_nodes:
                break

            # 更新alpha
            alpha = min_g

            # 剪枝所有g(t)=alpha的节点
            for node in min_g_nodes:
                node.clear_child()

            # 更新整棵树的R_T和leaf_count
            self._update_tree_metrics(T)

            # 保存当前子树和alpha
            k += 1
            self.pruned_trees.append(deepcopy(T))
            self.alphas.append(alpha)

            # 如果达到ccp_alpha阈值则停止
            if alpha > self.ccp_alpha:
                break

        # 步骤7: 使用验证集选择最优子树
        if val_df is not None:
            best_mse = float("inf")
            best_tree_idx = -1

            for i, tree in enumerate(self.pruned_trees):
                cur_mse = self.calculate_mse_from_node(val_df, label_col, tree)
                if cur_mse < best_mse:
                    best_mse = cur_mse
                    best_tree_idx = i

            # 选择最优子树
            self.root = self.pruned_trees[best_tree_idx]
            print(
                f"Selected subtree T_{best_tree_idx} with alpha={self.alphas[best_tree_idx]:.6f}, accuracy={best_mse:.4f}"
            )

    def _update_tree_metrics(self, node):
        if node.is_leaf():
            return

        total_R_T = 0.0
        total_leaf_count = 0
        for child in node.children.values():
            self._update_tree_metrics(child)
            total_R_T += child.R_T
            total_leaf_count += child.leaf_count

        node.R_T = total_R_T
        node.leaf_count = total_leaf_count

    def predict_row(self, row: pd.Series, node) -> str:
        q = Queue()
        q.put((node, 1.0))
        ans = []
        while not q.empty():
            current_node, current_weight = q.get()
            if current_node.is_leaf():
                ans.append(current_weight * current_node.value)
                continue

            if current_node.feature_name not in row.index:
                print(f"Warning: Feature {current_node.feature_name} not found in row.")
                ans.append(current_weight * current_node.value)
                continue

            feature_value = row[current_node.feature_name]
            if pd.notna(feature_value):
                if current_node.is_continuous:
                    if feature_value <= current_node.threshold:
                        child_node = current_node.children["<="]
                    else:
                        child_node = current_node.children[">"]
                    q.put((child_node, current_weight))
                else:
                    if feature_value == current_node.threshold:
                        child_node = current_node.children["<="]

                    else:
                        child_node = current_node.children[">"]
                    q.put((child_node, current_weight))
            else:
                for value, child_node in current_node.children.items():
                    q.put((child_node, current_weight * current_node.ratio[value]))

        # Combine results from all paths
        assert len(ans) > 0
        return sum(ans)

    def calculate_mse_from_node(self, df, label_col, node):
        if not df or df.empty:
            return 0.0
        preds = df.apply(lambda row: self.predict_row(row, node), axis=1)
        return ((preds - df[label_col]) ** 2).mean()

    def predict(self, df: pd.DataFrame) -> pd.Series:
        if self.root is None:
            raise ValueError("The model has not been fitted yet.")
        return df.apply(lambda row: self.predict_row(row, self.root), axis=1)

    def print_tree(self, node=None, indent: str = "", prefix: str = ""):
        """打印决策树结构"""
        if node is None:
            node = self.root
            print("Decision Tree:")
            print(f"Total leaves: {node.leaf_count}")

        # 叶子节点打印预测值
        if node.is_leaf():
            print(
                f"{indent}{prefix} Leaf: value={node.value:.4f} "
                f"(R={node.R:.2f}, samples={node.node_weight:.3f})"
            )
            return

        # 内部节点打印特征信息
        if node.is_continuous:
            print(
                f"{indent}{prefix} {node.feature_name} <= {node.threshold:.4f} "
                f"[R_T={node.R_T:.2f}, leaves={node.leaf_count}]"
            )
            # 递归打印子节点
            self.print_tree(node.children.get("<="), indent + "  ", "├── <=: ")
            self.print_tree(node.children.get(">"), indent + "  ", "└── >: ")
        else:
            print(
                f"{indent}{prefix} {node.feature_name} == {node.threshold} "
                f"[R_T={node.R_T:.2f}, leaves={node.leaf_count}]"
            )
            self.print_tree(node.children.get("<="), indent + "  ", "├── ==: ")
            self.print_tree(node.children.get(">"), indent + "  ", "└── !=: ")


if __name__ == "__main__":
    pass
相关推荐
我有一些感想……18 小时前
浅谈 BSGS(Baby-Step Giant-Step 大步小步)算法
c++·算法·数论·离散对数·bsgs
麦麦大数据18 小时前
F042 A星算法课程推荐(A*算法) | 课程知识图谱|课程推荐vue+flask+neo4j B/S架构前后端分离|课程知识图谱构造
vue.js·算法·知识图谱·neo4j·a星算法·路径推荐·课程推荐
丰锋ff18 小时前
英一2016年真题学习笔记
笔记·学习
Lynnxiaowen18 小时前
今天我们学习Linux架构keepalived实现LVS代理双击热备
linux·学习·架构·云计算
贝塔实验室18 小时前
LDPC 码的度分布
线性代数·算法·数学建模·fpga开发·硬件工程·信息与通信·信号处理
快手技术18 小时前
端到端短视频多目标排序机制框架 EMER 详解
算法
Wenhao.18 小时前
LeetCode LRU缓存
算法·leetcode·缓存·golang
京东零售技术19 小时前
告别 “盲买”!京东 AI 试穿 Oxygen Tryon:让服饰购物从“想象”到“所见即所得”
算法
小白菜又菜19 小时前
Leetcode 2273. Find Resultant Array After Removing Anagrams
算法·leetcode·职场和发展
milanyangbo19 小时前
谁生?谁死?从引用计数到可达性分析,洞悉GC的决策逻辑
java·服务器·开发语言·jvm·后端·算法·架构