[1]Wikipedia. 中華民國法律. from https://zh.wikipedia.org/zh-tw/%E4%B8%AD%E8%8F%AF%E6%B0%91%E5%9C%8B%E6%B3%95%E5%BE%8B
[2]中央研究院-詞庫小組. (1986). CHINESE KNOWLEDGE AND INFORMATION PROCESSING. from https://ckip.iis.sinica.edu.tw/
[3] 王安定(2016)。判決書之探勘分析與量刑迴歸模型之建立~以法院毒品判決書為例。未出版之碩士論文,臺北市立大學資訊科學系碩士在職專班,臺北市。[4]司法院. 民事訴訟. from https://www.judicial.gov.tw/tw/cp-1447-58182-7cb42-1.html
[5]司法院法學資料檢索系統()。裁判書查詢系統。 取自 https://law.judicial.gov.tw/FJUD/Default_AD.aspx
[6]全國法規資料庫()。刑事訴訟法 第 308 條。 取自 https://law.moj.gov.tw/LawClass/LawSingle.aspx?pcode=C0010001&flno=308
[7]全國法規資料庫()。法院組織法 第 83 條。 取自 https://law.moj.gov.tw/LawClass/LawSingle.aspx?pcode=A0010053&flno=83
[8]朱璟軒(2021)。一個基於深度學習之刑度預測模型—以臺灣地區竊盜案件為例。未出版之碩士論文,國立臺灣科技大學資訊工程系,台北市。[9]吳晨皓(2020)。BERT與GPT-2分別應用於刑事案件之罪名分類及判決書生成。未出版之碩士論文,國立高雄科技大學資訊管理系,高雄市。[10]林婷嫻. (2018). 斷開中文的鎖鍊!自然語言處理 (NLP)是什麼?. Retrieved from https://research.sinica.edu.tw/nlp-natural-language-processing-chinese-knowledge-information/
[11]紀幸辰(2018)。一個基於機器學習的醫療法判決書預測系統使用具文本相似性的法條分類決策樹。未出版之碩士論文,國立臺北大學資訊工程學系,新北市。[12]高點法商編委會(2022)。高點體系式分類六法:刑事法規(含法律倫理)):高點文化事業有限公司。
[13]曹錫璋(2021)。基於深度學習模型之判決書情境相似檢索技術之研究。未出版之碩士論文,國立中興大學資訊科學與工程學系所,台中市。[14]陳冠瑜(2015)。利用語意分析模型分析谷歌部落格搜尋引擎效能。未出版之碩士論文,國立東華大學資訊管理碩士學位學程,花蓮縣。[15]曾紀秀(2021)。酒駕犯罪判決書之大數據採礦分析。未出版之碩士論文,國立中正大學雲端計算與物聯網數位學習碩士在職專班,嘉義縣。[16]匯澤法律事務所. 刑事案件. from https://huitselaw.com/service/list/Criminal-case/
[17]楊舒婷(2020)。怎麼看懂判決書?判決書上都記載了什麼內容?(二)——刑事判決書。 取自 https://www.legis-pedia.com/article/lawABC/739
[18] 謝閎宇(2021)。基於深度學習模型之刑事判決書情境萃取研究。未出版之碩士論文,國立中興大學資訊科學與工程學系所,台中市。[19]謝德成(2021)。以機器學習為基礎之車禍致死案件精神損害賠償判決預測。未出版之碩士論文,國立暨南國際大學資訊工程學系,南投縣。[20]Bafna, P., Pramod, D., & Vaidya, A. (2016). Document clustering: TF-IDF approach. Paper presented at the 2016 International Conference on Electrical, Electronics, and Optimization Techniques (ICEEOT).
[21]Beltagy, I., Peters, M. E., & Cohan, A. J. a. p. a. (2020). Longformer: The long-document transformer.
[22]Bradley, A. P. J. P. r. (1997). The use of the area under the ROC curve in the evaluation of machine learning algorithms. 30(7), 1145-1159.
[23]Brand, M. J. L. a., & applications, i. (2006). Fast low-rank modifications of the thin singular value decomposition. 415(1), 20-30.
[24]Chen, T., Xu, B., Zhang, C., & Guestrin, C. J. a. p. a. (2016). Training deep nets with sublinear memory cost.
[25]Cui, Y., Che, W., Liu, T., Qin, B., Yang, Z. J. I. A. T. o. A., Speech,, & Processing, L. (2021). Pre-training with whole word masking for chinese bert. 29, 3504-3514.
[26]Cui, Y., Che, W., Liu, T., Qin, B., Wang, S., & Hu, G. J. a. p. a. (2020). Revisiting pre-trained models for Chinese natural language processing.
[27]Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. J. a. p. a. (2018). Bert: Pre-training of deep bidirectional transformers for language understanding.
[28]Giri, R. K., Gupta, S. C., & Gupta, U. K. (2021). An approach to detect offence in memes using natural language processing (NLP) and deep learning. Paper presented at the 2021 International Conference on Computer Communication and Informatics (ICCCI).
[29]Heydarian, M., Doyle, T. E., & Samavi, R. J. I. A. (2022). MLCM: Multi-label confusion matrix. 10, 19083-19095.
[30]Hofmann, T. (1999). Probabilistic latent semantic indexing. Paper presented at the Proceedings of the 22nd annual international ACM SIGIR conference on Research and development in information retrieval.
[31]Is there a method for calculating the best accuracy threshold for multi-label classification? (Lesson 3). (2019). from https://forums.fast.ai/t/is-there-a-method-for-calculating-the-best-accuracy-threshold-for-multi-label-classification-lesson-3/75330
[32]Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P., & Soricut, R. J. a. p. a. (2019). Albert: A lite bert for self-supervised learning of language representations.
[33]Li, P.-H., Fu, T.-J., & Ma, W.-Y. (2020). Why attention? Analyze BiLSTM deficiency and its remedies in the case of NER. Paper presented at the Proceedings of the AAAI Conference on Artificial Intelligence.
[34]Liddy, E. D. (2001). Natural language processing.
[35]Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., . . . Stoyanov, V. J. a. p. a. (2019). Roberta: A robustly optimized bert pretraining approach.
[36]Mihalcea, R., & Tarau, P. (2004). Textrank: Bringing order into text. Paper presented at the Proceedings of the 2004 conference on empirical methods in natural language processing.
[37]Mikolov, T., Chen, K., Corrado, G., & Dean, J. J. a. p. a. (2013). Efficient estimation of word representations in vector space.
[38]O'Shea, K., & Nash, R. J. a. p. a. (2015). An introduction to convolutional neural networks.
[39]OpenAI. (2022). Chat Generative Pre-trained Transformer,ChatGPT. from https://chat.openai.com/chat
[40]Seliya, N., Khoshgoftaar, T. M., & Van Hulse, J. (2009). A study on the relationships of classifier performance metrics. Paper presented at the 2009 21st IEEE international conference on tools with artificial intelligence.
[41]Sherstinsky, A. J. P. D. N. P. (2020). Fundamentals of recurrent neural network (RNN) and long short-term memory (LSTM) network. 404, 132306.
[42]Staudemeyer, R. C., & Morris, E. R. J. a. p. a. (2019). Understanding LSTM--a tutorial into long short-term memory recurrent neural networks.
[43]Tsoumakas, G., Katakis, I., & Vlahavas, I. (2006). A review of multi-label classification methods. Paper presented at the Proceedings of the 2nd ADBIS workshop on data mining and knowledge discovery (ADMKD 2006).
[44] Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., . . . Polosukhin, I. J. A. i. n. i. p. s. (2017). Attention is all you need. 30.
[45]Wang, Y., Gao, J., & Chen, J. (2020). Deep learning algorithm for judicial judgment prediction based on BERT. Paper presented at the 2020 5th International Conference on Computing, Communication and Security (ICCCS).
[46]Yang, M. (2020). ckiplab/albert-base-chinese. from https://huggingface.co/ckiplab/albert-base-chinese
[47]Yang, M. (2020). ckiplab/bert-base-chinese. from https://huggingface.co/ckiplab/bert-base-chinese