6 CONCLUSION
Our study motivates the use of opinion synopses
rather than long-length descriptions to predict issue
areas at scale. We analyzed qualitatively whether le-
gal is a domain-specific problem from an NLP tools
perspective, or a domain that could be generalized
based on representation and the task of interest. Fine-
tuned on our balanced dataset with the most dissim-
ilar splits, we showed that a sustainable generalized
language-model is more train-efficient and outper-
formed a model pretrained on a specialized legal do-
main.
Our results carve several avenues of future re-
search such as improve performance by removing
name entities from summaries, apply text simplifica-
tion to auto-generate opinion abstractions from long
documents, and expand our work to a broad class of
prediction tasks in legal studies. As it becomes in-
creasingly important to develop simple, efficient, and
reproducible domain-agnostic systems for neural text
processing, we hope our approach will help the NLP
community to further expand prediction analysis to
other humanity disciplines.
ACKNOWLEDGMENTS
We would like to thank the anonymous reviewers for
their insightful suggestions and feedback.
REFERENCES
Ashley, K. D. and Br
¨
uninghaus, S. (2009). Automatically
classifying case texts and predicting outcomes. Artifi-
cial Intelligence and Law, 17(2):125–165.
Chalkidis, I., Fergadiotis, E., Malakasiotis, P., Aletras, N.,
and Androutsopoulos, I. (2019). Extreme multi-label
legal text classification: A case study in EU legisla-
tion. In Natural Legal Language Processing Work-
shop, pages 78–87, Minneapolis, Minnesota.
Chalkidis, I., Fergadiotis, M., Malakasiotis, P., Aletras,
N., and Androutsopoulos, I. (2020). LEGAL-BERT:
The muppets straight out of law school. In Find-
ings of the Association for Computational Linguistics:
(EMNLP), pages 2898–2904, Online.
Chen, D. (2019). Judicial analytics and the great transfor-
mation of American law. Artificial Intelligence and
the Law, 27(1):15–42.
Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K.
(2019). BERT: Pre-training of deep bidirectional
transformers for language understanding. In North
American Chapter of the Association for Compu-
tational Linguistics: Human Language Technolo-
gies (NAACL), pages 4171–4186, Minneapolis, Min-
nesota.
Evans, M., McIntosh, W., Lin, J., and Cates, C. (2007).
Recounting the courts? applying automated content
analysis to enhance empirical legal research. Journal
of Empirical Legal Studies, 4(4).
Katz, D. M., Bommarito, M. J., and Blackman, J. (2017).
A general approach for predicting the behavior of the
supreme court of the united states. PLOS ONE, 12(4).
Kingma, D. P. and Ba, J. (2014). Adam: A method for
stochastic optimization. CoRR, abs/1412.6980. http:
//arxiv.org/abs/1412.6980.
Koreeda, Y. and Manning, C. D. (2021). Contractnli: A
dataset for document-level natural language inference
for contracts. Available at https://export.arxiv.org/abs/
2110.01799v1.
Lame, G. (2005). Using NLP Techniques to Identify Legal
Ontology Components: Concepts and Relations, page
169–184. Springer-Verlag, Berlin, Heidelberg.
Le, Q. and Mikolov, T. (2014). Distributed representations
of sentences and documents. In International Confer-
ence on Machine Learning (ICML), pages 1188–1196,
Bejing, China.
Lewis, M., Liu, Y., Goyal, N., Ghazvininejad, M., Mo-
hamed, A., Levy, O., Stoyanov, V., and Zettlemoyer,
L. (2020). BART: Denoising sequence-to-sequence
pre-training for natural language generation, trans-
lation, and comprehension. In Annual Meeting of
the Association for Computational Linguistics (ACL),
pages 7871–7880, Online.
Merchant, K. and Pande, Y. (2018). NLP based latent se-
mantic analysis for legal text summarization. In Ad-
vances in Computing, Communications and Informat-
ics (ICACCI), pages 1803–1807, Bangalore, India.
Nallapati, R. and Manning, C. D. (2008). Legal docket
classification: Where machine learning stumbles. In
Empirical Methods in Natural Language Processing
(EMNLP), pages 438–446, Honolulu, Hawaii.
Reimers, N. and Gurevych, I. (2019). Sentence-BERT: Sen-
tence embeddings using Siamese BERT-networks. In
Empirical Methods in Natural Language Processing
(EMNLP-IJCNLP), pages 3982–3992, Hong Kong,
China.
Rubner, Y., Tomasi, C., and Guibas, L. (1998). A metric for
distributions with applications to image databases. In
IEEE Conference on Computer Vision, pages 59–66.
Sanh, V., Debut, L., Chaumond, J., and Wolf, T. (2019).
Distilbert, a distilled version of BERT: smaller, faster,
cheaper and lighter. CoRR, abs/1910.01108. http://
arxiv.org/abs/1910.01108.
Sarkar, R., Ojha, A. K., Megaro, J., Mariano, J., Herard, V.,
and McCrae, J. P. (2021). Few-shot and zero-shot ap-
proaches to legal text classification: A case study in
the financial sector. In Natural Legal Language Pro-
cessing (NLLP), pages 102–106, Online.
Søgaard, A., Ebert, S., Bastings, J., and Filippova, K.
(2021). We need to talk about random splits. In Eu-
ropean Chapter of the Association for Computational
Linguistics (EACL), pages 1823–1832, Online.
Issue Area Discovery from Legal Opinion Summaries using Neural Text Processing
943