机器学习导论习题解答

在文档聚类中,通过考虑上下文,例如考虑像"cocktail party"与"party elections"中的词对,可以减少二义性。实现方法。

python 复制代码
import re
import numpy as np
import matplotlib.pyplot as plt
from nltk.corpus import stopwords
from nltk.tokenize import word_tokenize
from gensim.models import Word2Vec
from sklearn.feature_extraction.text import TfidfVectorizer
from sklearn.cluster import KMeans
from sklearn.decomposition import PCA
from sklearn.metrics import silhouette_score
# ========== 新增:配置Matplotlib支持中文 ==========
plt.rcParams['font.sans-serif'] = ['Microsoft YaHei']  # 指定默认字体(微软雅黑)
plt.rcParams['axes.unicode_minus'] = False  # 解决负号显示为方块的问题
# ===============================================

# ---------------------- 1. 初始化配置与数据准备 ----------------------
# 下载nltk停用词(首次运行需要)
import nltk
nltk.download('stopwords')
nltk.download('punkt')

# 定义示例文档集(包含"party"的不同上下文)
documents = [
    "We had a great cocktail party at the beach with friends",
    "The cocktail party had delicious drinks and music",
    "The political party won the elections by a large margin",
    "Party elections are held every four years in the country",
    "I attended a birthday party with family last weekend",
    "The ruling party announced new policies after the elections",
    "Beach party with cocktails is my favorite summer activity",
    "Elections campaign for the party started last month"
]

# 定义停用词表
stop_words = set(stopwords.words('english'))

# ---------------------- 2. 文本预处理函数 ----------------------
def preprocess_text(text):
    """
    文本预处理:小写化、去除特殊字符、分词、去停用词
    """
    # 小写化
    text = text.lower()
    # 去除非字母字符(保留空格)
    text = re.sub(r'[^a-zA-Z\s]', '', text)
    # 分词
    tokens = word_tokenize(text)
    # 去停用词和空字符串
    tokens = [token for token in tokens if token not in stop_words and token.strip() != '']
    return tokens

# 对所有文档进行预处理
processed_docs = [preprocess_text(doc) for doc in documents]
# 为TF-IDF准备拼接后的文本(需要字符串格式)
tfidf_docs = [' '.join(tokens) for tokens in processed_docs]

# ---------------------- 3. 捕捉局部上下文:2-gram TF-IDF ----------------------
# 初始化TF-IDF向量化器(包含1-gram和2-gram,捕捉词对上下文)
tfidf_vectorizer = TfidfVectorizer(ngram_range=(1, 2))
# 生成2-gram TF-IDF特征
tfidf_features = tfidf_vectorizer.fit_transform(tfidf_docs).toarray()
print(f"2-gram TF-IDF特征维度: {tfidf_features.shape}")

# ---------------------- 4. 捕捉全局语义上下文:Word2Vec ----------------------
# 训练Word2Vec模型(捕捉词的全局上下文语义)
w2v_model = Word2Vec(
    sentences=processed_docs,  # 训练语料
    vector_size=100,           # 词嵌入维度
    window=5,                  # 上下文窗口大小(前后5个词)
    min_count=1,               # 最小词频(保留所有词)
    workers=4                  # 并行线程数
)

def get_doc_vector(tokens, model, vector_size=100):
    """
    将文档转换为Word2Vec向量(所有词向量的平均值)
    """
    doc_vectors = []
    for token in tokens:
        if token in model.wv:
            doc_vectors.append(model.wv[token])
    if not doc_vectors:
        # 如果文档无有效词,返回全零向量
        return np.zeros(vector_size)
    # 取平均值作为文档向量
    return np.mean(doc_vectors, axis=0)

# 生成所有文档的Word2Vec向量
w2v_doc_vectors = np.array([get_doc_vector(tokens, w2v_model) for tokens in processed_docs])
print(f"Word2Vec文档向量维度: {w2v_doc_vectors.shape}")

# ---------------------- 5. 特征融合:合并上下文特征 ----------------------
# 拼接TF-IDF(局部上下文)和Word2Vec(全局上下文)特征
combined_features = np.hstack((tfidf_features, w2v_doc_vectors))
print(f"融合后特征维度: {combined_features.shape}")

# ---------------------- 6. 文档聚类(KMeans) ----------------------
# 设定聚类数(根据示例数据,分为2类:聚会/政党)
n_clusters = 2
# 初始化KMeans聚类器
kmeans = KMeans(n_clusters=n_clusters, random_state=42)
# 执行聚类
cluster_labels = kmeans.fit_predict(combined_features)

# 输出聚类结果
print("\n=== 聚类结果 ===")
for i, (doc, label) in enumerate(zip(documents, cluster_labels)):
    print(f"文档{i+1}: {doc[:50]}... | 聚类标签: {label}")

# 计算轮廓系数(评估聚类效果,越接近1越好)
sil_score = silhouette_score(combined_features, cluster_labels)
print(f"\n聚类轮廓系数: {sil_score:.4f}")

# ---------------------- 7. 聚类结果可视化(PCA降维) ----------------------
# PCA降维到2维
pca = PCA(n_components=2)
reduced_features = pca.fit_transform(combined_features)

# 绘制散点图
plt.figure(figsize=(10, 6))
colors = ['red', 'blue']
# labels = ['聚会相关(party=聚会)', '政党相关(party=政党)']
labels = ['Party (social gathering)', 'Party (political party)']
for i in range(n_clusters):
    mask = cluster_labels == i
    plt.scatter(
        reduced_features[mask, 0], 
        reduced_features[mask, 1],
        c=colors[i],
        label=labels[i],
        alpha=0.7
    )

# 添加文档编号标注
for i, (x, y) in enumerate(reduced_features):
    plt.annotate(f"Doc{i+1}", (x, y), fontsize=8)

# plt.title('文档聚类结果(考虑上下文减少party二义性)')
# 修改为英文
plt.title('Document Clustering Results (Reducing "party" Ambiguity with Context)')

plt.xlabel('PCA维度1')
plt.ylabel('PCA维度2')
plt.legend()
plt.grid(True, alpha=0.3)
plt.show()

# ---------------------- 8. 验证二义性减少效果:对比"party"的不同上下文向量 ----------------------
print("\n=== 验证party的上下文向量差异 ===")
# 提取"party"在不同上下文的词向量
party_cocktail = w2v_model.wv['party'] - w2v_model.wv['cocktail']
party_elections = w2v_model.wv['party'] - w2v_model.wv['elections']
# 计算余弦相似度(值越小,说明上下文差异越大)
from sklearn.metrics.pairwise import cosine_similarity
similarity = cosine_similarity([party_cocktail], [party_elections])[0][0]
print(f"'party+cocktail'与'party+elections'的向量相似度: {similarity:.4f}")
相关推荐
独好紫罗兰5 小时前
对python的再认识-基于数据结构进行-a004-列表-实用事务
开发语言·数据结构·python
ZH15455891315 小时前
Flutter for OpenHarmony Python学习助手实战:模块与包管理的实现
python·学习·flutter
不穿格子的程序员5 小时前
从零开始刷算法——贪心篇1:跳跃游戏1 + 跳跃游戏2
算法·游戏·贪心
人工智能AI技术5 小时前
GitHub Copilot免费替代方案:大学生如何用CodeGeeX+通义灵码搭建AI编程环境
人工智能
Chunyyyen5 小时前
【第三十四周】视觉RAG01
人工智能·chatgpt
大江东去浪淘尽千古风流人物5 小时前
【SLAM新范式】几何主导=》几何+学习+语义+高效表示的融合
深度学习·算法·slam
是枚小菜鸡儿吖5 小时前
CANN 算子开发黑科技:AI 自动生成高性能 Kernel 代码
人工智能·科技
choke2335 小时前
[特殊字符] Python异常处理
开发语言·python
hqyjzsb5 小时前
盲目用AI提效?当心陷入“工具奴”陷阱,效率不增反降
人工智能·学习·职场和发展·创业创新·学习方法·业界资讯·远程工作