“Sunwei”的版本间的差异

来自南京大学IIP
跳转至: 导航搜索
   更改可见性
Lhy讨论 | 贡献
   更改可见性
第3行: 第3行:
 
|-
 
|-
 
|  
 
|  
M.Sc. Student @ IIP&nbsp;Group<br/> Department of Computer Science and Technology<br/> Nanjing University
+
<span style="font-size:larger;">M.Sc. Student @ IIP&nbsp;Group<br/> Department of Computer Science and Technology<br/> Nanjing University</span>
  
Email: weisun_@outlook.com
+
<span style="font-size:larger;">Email: weisun_@outlook.com</span>
  
 
|-
 
|-
第12行: 第12行:
 
|}
 
|}
  
<span style="color:#3498db"><span style="font-size:larger">'''Supervisor'''</span></span>
+
<span style="font-size:larger;"><span style="color:#3498db">'''Supervisor'''</span></span>
  
 
----
 
----
  
*Professor Jun-Yuan Xie  
+
*<span style="font-size:larger;">Professor Jun-Yuan Xie</span>
  
<span style="font-size:larger"><span style="color:#3498db">'''Biography'''</span></span>
+
<span style="font-size:larger;"><span style="color:#3498db">'''Biography'''</span></span>
  
 
----
 
----
  
*I&nbsp;received my B.Sc. degree in&nbsp;of Soochow University in June 2017. In the same year, I was admitted to study for a Master degree in Nanjing University&nbsp;without entrance examination.&nbsp; Currently I am a second&nbsp;year M.Sc. student of Department of Computer Science and Technology&nbsp;in Nanjing University&nbsp;and a member of IIP&nbsp;Group, led by professor Jun-Yuan Xie and&nbsp; Chong-Jun Wang.  
+
*<span style="font-size:larger;">I&nbsp;received my B.Sc. degree in&nbsp;of Soochow University in June 2017. In the same year, I was admitted to study for a Master degree in Nanjing University&nbsp;without entrance examination.&nbsp; Currently I am a second&nbsp;year M.Sc. student of Department of Computer Science and Technology&nbsp;in Nanjing University&nbsp;and a member of IIP&nbsp;Group, led by professor Jun-Yuan Xie and&nbsp; Chong-Jun Wang.</span>
  
<span style="font-size:larger"><span style="color:#3498db">'''Research Interest'''</span></span>
+
<span style="font-size:larger;"><span style="color:#3498db">'''Research Interest'''</span></span>
  
&nbsp;'''Machine Learning &&nbsp;Multi-label Learning'''.&nbsp;
+
<span style="font-size:larger;">&nbsp;'''Machine Learning &&nbsp;Multi-label Learning'''.&nbsp;</span>
  
*On '''Multi-label Text Classification (MLTC),''' text features can be regarded as '''detailed description of documents''' and label sets can be '''a summarization of documents'''. '''Hybrid Topics''' from text features and label sets by LDA (a method of '''topic model''') can effectively mine global label correlations and deeper&nbsp;features. Meanwhile, a pair including topics and labels can mitigate the imbalanced problem of labels.  
+
*<span style="font-size:larger;">On '''Multi-label Text Classification (MLTC),''' text features can be regarded as '''detailed description of documents''' and label sets can be '''a summarization of documents'''. '''Hybrid Topics''' from text features and label sets by LDA (a method of '''topic model''') can effectively mine global label correlations and deeper&nbsp;features. Meanwhile, a pair including topics and labels can mitigate the imbalanced problem of labels.</span>
*Deep learning For multi-label text classification. We utilize '''dilated convolution''' to obtain the '''semantic understanding''' of the text and design a hybrid '''attention mechansim''' for '''different labels''' (Specifically, each label should attend to most relevant textual contents).&nbsp; Firstly,&nbsp;we initialize trainable label embeddings. Then After obationing word-level information based on Bi-LSTM, we&nbsp;get semantic understanding of texts&nbsp;based on word-level information by dilated convolution. Finally,&nbsp;we design a hybrid attention for different labels based on label embeddings.&nbsp; Besides, we add '''label cooccurrence matrix into loss function '''to guide the whole network to learn and achieve good results.&nbsp;&nbsp;  
+
*<span style="font-size:larger;">Deep learning For multi-label text classification. We utilize '''dilated convolution''' to obtain the '''semantic understanding''' of the text and design a hybrid '''attention mechansim''' for '''different labels''' (Specifically, each label should attend to most relevant textual contents).&nbsp; Firstly,&nbsp;we initialize trainable label embeddings. Then After obationing word-level information based on Bi-LSTM, we&nbsp;get semantic understanding of texts&nbsp;based on word-level information by dilated convolution. Finally,&nbsp;we design a hybrid attention for different labels based on label embeddings.&nbsp; Besides, we add '''label cooccurrence matrix into loss function '''to guide the whole network to learn and achieve good results.&nbsp;&nbsp;</span>
*'''GCN (Graph Convolution Network) '''can be used to exploit more complex label correlations on Image Multi-label Learning.  
+
*<span style="font-size:larger;">'''GCN (Graph Convolution Network) '''can be used to exploit more complex label correlations on Image Multi-label Learning.</span>
  
<font color="#3498db"><span style="font-size:15.6px">'''Publications'''</span></font>
+
<span style="font-size:larger;"><font color="#3498db">'''Publications'''</font></span>
  
 
----
 
----
  
*Ran X., Pan Y., Sun W. and&nbsp;Wang C.. Learn to Select via Hierarchical Gate Mechanism for Aspect-Based Sentiment Analysis.&nbsp;In&nbsp;''Proceedings of IJCAL 2019.''  
+
*<span style="font-size:larger;">Ran X., Pan Y., Sun W. and&nbsp;Wang C.. Learn to Select via Hierarchical Gate Mechanism for Aspect-Based Sentiment Analysis.&nbsp;In&nbsp;''Proceedings of IJCAL 2019.''</span>
*Sun W., Wang C..Multi-label Classification: Select Distinct Semantic Understanding for Different Labels[C]. In&nbsp;''Proceedings of&nbsp;ApWeb-WAIM 2019.''  
+
*<span style="font-size:larger;">Sun W., Wang C..Multi-label Classification: Select Distinct Semantic Understanding for Different Labels[C]. In&nbsp;''Proceedings of&nbsp;ApWeb-WAIM 2019.''</span>
*Sun W., Ran X.. Luo X., and Wang C..An Efficient Framework by Topic Model for Multi-label Text Classification. In&nbsp;''Proceedings of IJCNN 2019.''  
+
*<span style="font-size:larger;">Sun W., Ran X.. Luo X., and Wang C..An Efficient Framework by Topic Model for Multi-label Text Classification. In&nbsp;''Proceedings of IJCNN 2019.''</span>
*Xu Y., Ran X.. Sun W., Luo X.&nbsp;and Wang C..Gated Neural Network with Regularized Loss for Multi-label Text Classification. In&nbsp;''Proceedings of IJCNN 2019.''  
+
*<span style="font-size:larger;">Xu Y., Ran X.. Sun W., Luo X.&nbsp;and Wang C..Gated Neural Network with Regularized Loss for Multi-label Text Classification. In&nbsp;''Proceedings of IJCNN 2019.''</span>
*Ran X., Pan Y., Sun W. and&nbsp;Wang C.. Modeling More Globally: A Hierarchical Attention Network via Multi-Task Learning for Aspect-Based Sentiment Analysis. In&nbsp;''Proceedings of DASFAA 2019'', Chiang Mai, Thailand, Apr. 22-25,&nbsp;2019: 505-509.  
+
*<span style="font-size:larger;">Ran X., Pan Y., Sun W. and&nbsp;Wang C.. Modeling More Globally: A Hierarchical Attention Network via Multi-Task Learning for Aspect-Based Sentiment Analysis. In&nbsp;''Proceedings of DASFAA 2019'', Chiang Mai, Thailand, Apr. 22-25,&nbsp;2019: 505-509.</span>
  
<span style="font-size:larger"><span style="color:#3498db">'''Resources'''</span></span>
+
<span style="font-size:larger;"><span style="color:#3498db">'''Resources'''</span></span>
  
 
----
 
----
  
*[http://manikvarma.org/downloads/XC/XMLRepository.html Extreme Classification Repository]: for large-scale multi-label datasets and off-the-shelf eXtreme Multi-Label Learning (XML) solvers.  
+
*<span style="font-size:larger;">[http://manikvarma.org/downloads/XC/XMLRepository.html Extreme Classification Repository]: for large-scale multi-label datasets and off-the-shelf eXtreme Multi-Label Learning (XML) solvers.</span>
*[http://mulan.sourceforge.net/datasets-mlc.html Mulan Multi-Label Learning Datasets]: regular/traditional multi-label learning datasets.  
+
*<span style="font-size:larger;">[http://mulan.sourceforge.net/datasets-mlc.html Mulan Multi-Label Learning Datasets]: regular/traditional multi-label learning datasets.</span>
*[https://github.com/XSilverBullet/Multi-label-Paper Related Work]: This page categorizes a list of works of my interest, mainly in Multi-Label Learning.  
+
*<span style="font-size:larger;">[https://github.com/XSilverBullet/Multi-label-Paper Related Work]: This page categorizes a list of works of my interest, mainly in Multi-Label Learning.</span>
  
<span style="font-size:larger"><span style="color:#3498db">'''Rewards or Honors'''</span></span>
+
<span style="font-size:larger;"><span style="color:#3498db">'''Rewards or Honors'''</span></span>
  
*Second-Class Academic Scholarship, 2018-2019  
+
*<span style="font-size:larger;">Second-Class Academic Scholarship, 2018-2019</span>
*First-Class Academic Scholarship, 2017-2018  
+
*<span style="font-size:larger;">First-Class Academic Scholarship, 2017-2018</span>
*Outstanding Graduate Student, 2017.06  
+
*<span style="font-size:larger;">Outstanding Graduate Student, 2017.06</span>
*CCF Excellent University Student, 2016.10  
+
*<span style="font-size:larger;">CCF Excellent University Student, 2016.10</span>
*National Scholarship, 2015.11
+
*<span style="font-size:larger;">National Scholarship, 2015.11</span>

2019年5月26日 (日) 23:34的版本

M.Sc. Student @ IIP Group
Department of Computer Science and Technology
Nanjing University

Email: weisun_@outlook.com

   

Supervisor


  • Professor Jun-Yuan Xie

Biography


  • I received my B.Sc. degree in of Soochow University in June 2017. In the same year, I was admitted to study for a Master degree in Nanjing University without entrance examination.  Currently I am a second year M.Sc. student of Department of Computer Science and Technology in Nanjing University and a member of IIP Group, led by professor Jun-Yuan Xie and  Chong-Jun Wang.

Research Interest

 Machine Learning & Multi-label Learning

  • On Multi-label Text Classification (MLTC), text features can be regarded as detailed description of documents and label sets can be a summarization of documents. Hybrid Topics from text features and label sets by LDA (a method of topic model) can effectively mine global label correlations and deeper features. Meanwhile, a pair including topics and labels can mitigate the imbalanced problem of labels.
  • Deep learning For multi-label text classification. We utilize dilated convolution to obtain the semantic understanding of the text and design a hybrid attention mechansim for different labels (Specifically, each label should attend to most relevant textual contents).  Firstly, we initialize trainable label embeddings. Then After obationing word-level information based on Bi-LSTM, we get semantic understanding of texts based on word-level information by dilated convolution. Finally, we design a hybrid attention for different labels based on label embeddings.  Besides, we add label cooccurrence matrix into loss function to guide the whole network to learn and achieve good results.  
  • GCN (Graph Convolution Network) can be used to exploit more complex label correlations on Image Multi-label Learning.

Publications


  • Ran X., Pan Y., Sun W. and Wang C.. Learn to Select via Hierarchical Gate Mechanism for Aspect-Based Sentiment Analysis. In Proceedings of IJCAL 2019.
  • Sun W., Wang C..Multi-label Classification: Select Distinct Semantic Understanding for Different Labels[C]. In Proceedings of ApWeb-WAIM 2019.
  • Sun W., Ran X.. Luo X., and Wang C..An Efficient Framework by Topic Model for Multi-label Text Classification. In Proceedings of IJCNN 2019.
  • Xu Y., Ran X.. Sun W., Luo X. and Wang C..Gated Neural Network with Regularized Loss for Multi-label Text Classification. In Proceedings of IJCNN 2019.
  • Ran X., Pan Y., Sun W. and Wang C.. Modeling More Globally: A Hierarchical Attention Network via Multi-Task Learning for Aspect-Based Sentiment Analysis. In Proceedings of DASFAA 2019, Chiang Mai, Thailand, Apr. 22-25, 2019: 505-509.

Resources


Rewards or Honors

  • Second-Class Academic Scholarship, 2018-2019
  • First-Class Academic Scholarship, 2017-2018
  • Outstanding Graduate Student, 2017.06
  • CCF Excellent University Student, 2016.10
  • National Scholarship, 2015.11