| |||||||||||||||
Big Learning 2011 : NIPS 2011 Workshop on Algorithms, Systems, and Tools for Learning at Scale | |||||||||||||||
Link: http://www.biglearn.org | |||||||||||||||
| |||||||||||||||
Call For Papers | |||||||||||||||
Big Learning: Algorithms, Systems, and Tools for Learning at Scale
NIPS 2011 Workshop (http://www.biglearn.org) Submissions are solicited for a two day workshop December 16-17 in Sierra Nevada, Spain. This workshop will address tools, algorithms, systems, hardware, and real-world problem domains related to large-scale machine learning (“Big Learning”). The Big Learning setting has attracted intense interest with active research spanning diverse fields including machine learning, databases, parallel and distributed systems, parallel architectures, and programming languages and abstractions. This workshop will bring together experts across these diverse communities to discuss recent progress, share tools and software, identify pressing new challenges, and to exchange new ideas. Topics of interest include (but are not limited to): * Hardware Accelerated Learning: Practicality and performance of specialized high-performance hardware (e.g. GPUs, FPGAs, ASIC) for machine learning applications. * Applications of Big Learning: Practical application case studies; insights on end-users, typical data workflow patterns, common data characteristics (stream or batch); trade-offs between labeling strategies (e.g., curated or crowd-sourced); challenges of real-world system building. * Tools, Software, & Systems: Languages and libraries for large-scale parallel or distributed learning. Preference will be given to approaches and systems that leverage cloud computing (e.g. Hadoop, DryadLINQ, EC2, Azure), scalable storage (e.g. RDBMs, NoSQL, graph databases), and/or specialized hardware (e.g. GPU, Multicore, FPGA, ASIC). * Models & Algorithms: Applicability of different learning techniques in different situations (e.g., simple statistics vs. large structured models); parallel acceleration of computationally intensive learning and inference; evaluation methodology; trade-offs between performance and engineering complexity; principled methods for dealing with large number of features; We suggest keeping the paper under 4 pages (NOT including references) in the NIPS latex style. For projects that require more room for descriptions, we encourage the authors to include details of the work as appendix and/or other supplementary materials. Relevant work previously presented in non-machine-learning conferences is strongly encouraged. Exciting work that was recently presented is allowed, provided that the extended abstract mentions this explicitly. Submission Deadline: October 7th, 2011. Please refer to the website for detailed submission instructions: http://biglearn.org/index.php/AuthorInfo |
|