@@ -51,17 +51,17 @@ def create_epub():
5151 # Define chapters in order
5252 chapters = [
5353 {'title' : 'README' , 'file' : 'docs/readme.xhtml' , 'id' : 'readme' },
54- {'title' : '前言' , 'file' : 'docs/preface.xhtml' , 'id' : 'preface' },
55- {'title' : '第一章 NLP基础概念' , 'file' : 'docs/chapter1.xhtml' , 'id' : 'chapter1' },
56- {'title' : '第二章 Transformer架构' , 'file' : 'docs/chapter2_transformer_architecture.xhtml' , 'id' : 'chapter2' },
57- {'title' : '第三章 预训练语言模型' , 'file' : 'docs/chapter3_pretrained_language_models.xhtml' , 'id' : 'chapter3' },
58- {'title' : '第四章 大语言模型' , 'file' : 'docs/chapter4/chapter4_large_language_models.xhtml' , 'id' : 'chapter4' },
59- {'title' : '第五章 动手搭建大模型' , 'file' : 'docs/chapter5_building_large_models.xhtml' , 'id' : 'chapter5' },
60- {'title' : '第六章 大模型训练流程实践' , 'file' : 'docs/chapter6/chapter6_training_pipeline.xhtml' , 'id' : 'chapter6' },
61- {'title' : '4.1 偏好对齐' , 'file' : 'docs/chapter6/chapter6_preference_alignment.xhtml' , 'id' : 'preference_alignment' },
62- {'title' : '第七章 大模型应用' , 'file' : 'docs/chapter7/chapter7_applications.xhtml' , 'id' : 'chapter7' },
63- {'title' : 'Transformer架构详解' , 'file' : 'Extra-Chapter/transformer-architecture/transformer-architecture.xhtml' , 'id' : 'transformer-architecture' },
64- {'title' : '为什么微调小型大语言模型' , 'file' : 'Extra-Chapter/why-fine-tune-small-large-language-models/why-fine-tune-small-large-language-models.xhtml' , 'id' : 'why-fine-tune' },
54+ # {'title': '前言', 'file': 'docs/preface.xhtml', 'id': 'preface'},
55+ # {'title': '第一章 NLP基础概念', 'file': 'docs/chapter1.xhtml', 'id': 'chapter1'},
56+ # {'title': '第二章 Transformer架构', 'file': 'docs/chapter2_transformer_architecture.xhtml', 'id': 'chapter2'},
57+ # {'title': '第三章 预训练语言模型', 'file': 'docs/chapter3_pretrained_language_models.xhtml', 'id': 'chapter3'},
58+ # {'title': '第四章 大语言模型', 'file': 'docs/chapter4/chapter4_large_language_models.xhtml', 'id': 'chapter4'},
59+ # {'title': '第五章 动手搭建大模型', 'file': 'docs/chapter5_building_large_models.xhtml', 'id': 'chapter5'},
60+ # {'title': '第六章 大模型训练流程实践', 'file': 'docs/chapter6/chapter6_training_pipeline.xhtml', 'id': 'chapter6'},
61+ # {'title': '4.1 偏好对齐', 'file': 'docs/chapter6/chapter6_preference_alignment.xhtml', 'id': 'preference_alignment'},
62+ # {'title': '第七章 大模型应用', 'file': 'docs/chapter7/chapter7_applications.xhtml', 'id': 'chapter7'},
63+ # {'title': 'Transformer架构详解', 'file': 'Extra-Chapter/transformer-architecture/transformer-architecture.xhtml', 'id': 'transformer-architecture'},
64+ # {'title': '为什么微调小型大语言模型', 'file': 'Extra-Chapter/why-fine-tune-small-large-language-models/why-fine-tune-small-large-language-models.xhtml', 'id': 'why-fine-tune'},
6565 ]
6666
6767 epub_chapters = []
0 commit comments