教育资源为主的文档平台

当前位置: 查字典文档网> 所有文档分类> 外语学习> 英语学习> The big challenges of big data

The big challenges of big data

上传者:梁列全
|
上传时间:2015-05-04
|
次下载

The big challenges of big data

THE BIG CHALLENGES As they grapple with increasingly large data sets, biologists and computer scientists uncork new bottlenecks.EMBL–EBITECHNOLOGY FEATURE

Extremely powerful computers are needed to help biologists to handle big-data traffic jams.BY VIVIEN MARX

Biologists are joining the big-data club. With the advent of high-throughput genomics, life scientists are starting to grapple with massive data sets, encountering challenges with handling, processing and mov-ing information that were once the domain of astronomers and high-energy physicists1. With every passing year, they turn more often to big data to probe everything from the regulation of genes and the evolution of genomes to why coastal algae bloom, what microbes dwell where in human body cavities and how the genetic make-up of different can-cers influences how cancer patients fare2. The European Bioinformatics Institute (EBI) in Hinxton, UK, part of the European Molecular Biology Laboratory and one of the world’s larg-est biology-data repositories, currently stores 20 petabytes (1 petabyte is 1015 bytes) of data and back-ups about genes, proteins and small molecules. Genomic data account for 2 peta-bytes of that, a number that more than doubles every year3 (see ‘Data explosion’). This data pile is just one-tenth the size of the data store at CERN, Europe’s particle-physics laboratory near Geneva, Switzerland. Every

© 2013 Macmillan Publishers Limited. All rights reservedyear, particle-collision events in CERN’s Large Hadron Collider generate around 15 petabytes of data — the equivalent of about 4 million high-definition feature-length films. But the EBI and institutes like it face similar data-wrangling challenges to those at CERN, says Ewan Birney, associate director of the EBI. He and his colleagues now regularly meet with organizations such as CERN and the European Space Agency (ESA) in Paris to swap lessons about data storage, analysis and sharing.All labs need to manipulate data to yield

内容需要下载文档才能查看

research answers. As prices drop for high-throughput instruments such as automated 13 JUNE 2013 | VOL 498 | NATURE | 255

genome sequencers, small biology labs can

内容需要下载文档才能查看

become big-data generators. And even labs

内容需要下载文档才能查看

without such instruments can become big-data users by accessing terabytes (1012 bytes) of data from public repositories at the EBI or the US National Center for Biotechnology Information in Bethesda, Maryland. Each day last year, the EBI received about 9 mil-lion online requests to query its data, a 60% increase over 2011.Biology data mining has challenges all of its own, says Birney. Biological data are much more heterogeneous than those in physics. They stem from a wide range of experiments that spit out many types of information, such as genetic sequences, interactions of proteins or findings in medical records. The complexity is daunting, says Lawrence Hunter, a compu-tational biologist at the University of Colo-rado Denver. “Getting the most from the data requires interpreting them in light of all the relevant prior knowledge,” he says.That means scientists have to store large data sets, and analyse, compare and share them — not simple tasks. Even a single sequenced human genome is around 140 gigabytes in size. Comparing human genomes takes more than a personal computer and online file-sharing applications such as DropBox.In an ongoing study, Arend Sidow, a com-putational biologist at Stanford University in California, and his team are looking at specific changes in the genome sequences of tumours from people with breast cancer. They wanted to compare their data with the thousands of other published breast-cancer genomes and look for similar patterns in the scores of dif-ferent cancer types. But that is a tall order: downloading the data is time-consuming, and researchers must be sure that their com-putational infrastructure and software tools are up to the task. “If I could, I would routinely look at all sequenced cancer genomes,” says Sidow. “With the current infrastructure, that’s impossible.” In 2009, Sidow co-founded a company called DNAnexus in Mountain View, Califor-nia, to help with large-scale genetic analyses. Numerous other commercial and academic Republic, for example, might have an idea about how to reprocess cancer data to help the hunt for cancer drugs. If he or she lacks the computational equipment to develop it, he or she might not even try. But access to a high-powered cloud allows “ideas to come from any place”, says Birney. Even at the EBI, many scientists access databases and software tools on the Web and through clouds. “People rarely work on straight hardware anymore,” says Birney. One heavily used resource is the Ensembl Genome Browser, run jointly by the EBI and the Well-come Trust Sanger Institute in Hinxton. Life scientists use it to search through, down-load and analyse genomes from armadillo to zebrafish. The main Ensembl site is based on hardware in the United Kingdom, but when users in the United States and Japan had dif-Andreas Sundquist says amounts of data are now ficulty accessing the data quickly, the EBI larger than the tools used to analyse them.resolved the bottleneck by hosting mirror sites at three of the many remote data centres efforts also address the infrastructure needs of that are part of Amazon Web Services’ Elastic big-data biology. With the new types of data Compute Cloud (EC2). Amazon’s data centres traffic jam honking for attention, “we now have are geographically closer to the users than the non-trivial engineering problems”, says Birney, EBI base, giving researchers quicker access to the information they need.LIFE OF THE DATA-RICHMore clouds are coming. Together with Storing and interpreting big data takes both CERN and ESA, the EBI is building a cloud-real and virtual bricks and mortar. On the EBI based infrastructure called Helix Nebula campus, for example, construction is under — The Science Cloud. Also involved are infor-way to house the technical command centre mation-technology “If I could, I of ELIXIR, a project to help scientists across companies such would routinely Europe safeguard and share their data, and to as Atos in Bezons, support existing resources such as databases look at all France; CGI in Mon-and computing facilities in individual coun-

内容需要下载文档才能查看

sequenced treal, Canada; SixSq tries. Whereas CERN has one super collider cancer genomes. in Geneva; and T-Sys-producing data in one location, biological With the current tems in Frankfurt, research generating high volumes of data is infrastructure, Germany. distributed across many labs — highlighting that’s Cloud computing is the need to share resources. particularly attractive impossible.”Much of the construction in big-data biol-in an era of reduced ogy is virtual, focused on cloud computing research funding, says Hunter, because cloud — in which data and software are situated in users do not need to finance or maintain hard-huge, off-site centres that users can access on ware. In addition to academic cloud projects, demand, so that they do not need to buy their scientists can choose from many commercial own hardware and maintain it on site. Labs that providers, such as Rackspace, headquartered do have their own hardware can supplement it in San Antonio, Texas, or VMware in Palo with the cloud and use both as needed. They Alto, California, as well as larger companies can create virtual spaces for data, software and including Amazon, headquartered in Seattle, results that anyone can access, or they can lock Washington, IBM in Armonk, New York, or the spaces up behind a firewall so that only a Microsoft in Redmond, Washington.select group of collaborators can get to them. Working with the CSC — IT Center for Sci-BIG-DATA PARKING ence in Espoo, Finland, a government-run Clouds are a solution, but they also throw high-performance computing centre, the up fresh challenges. Ironically, their prolif-EBI is developing Embassy Cloud, a cloud-eration can cause a bottleneck if data end computing component for ELIXIR that offers up parked on several clouds and thus still secure data-analysis environments and is cur-need to be moved to be shared. And using rently in its pilot phase. External organizations clouds means entrusting valuable data to a can, for example, run data-driven experiments distant service provider who may be subject in the EBI’s computational environment, close to power outages or other disruptions. “I use to the data they need. They can also download cloud services for many things, but always data to compare with their own. keep a local copy of scientifically important The idea is to broaden access to computing data and software,” says Hunter. Scientists power, says Birney. A researcher in the Czech experiment with different constellations to

13 JUNE 2013 | VOL 498 | NATURE | 257DNANEXUSSOURCE: EMBL–EBI© 2013 Macmillan Publishers Limited. All rights reserved

suit their needs and trust levels. Most researchers tend to download remote data to local hardware for analysis. But this method is “backward”, says Andreas Sundquist, chief technology officer of DNAnexus. “The data are so much larger than the tools, it makes no sense to be doing that.” The alternative is to use the cloud for both data storage and com-puting. If the data are on a cloud, researchers can harness both the computing power and the tools that they need online, without the need to move data and software (see ‘Head in the clouds’). “There’s no reason to move data out-side the cloud. You can do analysis right there,” says Sundquist. Everything required is avail-able “to the clever people with the clever ideas”, regardless of their local computing resources, says Birney. Various academic and commercial ventures are engineering ways to bring data and analysis tools together — and as they build, they have to address the continued data growth. Xing Xu, director of cloud computing at BGI (formerly the Beijing Genomics Institute) in Shenzen, China, knows that challenge well. BGI is one of the largest producers of genomic data in the world, with 157 genome sequencing instru-ments working around the clock on samples from people, plants, animals and microbes. Each day, it generates 6 terabytes of genomic data. Every instrument can decode one human genome per week, an effort that used to take months or years and many staff. SOURCE: ASPERAHEAD IN THE CLOUDSIn cloud computing, large data sets are processed on remote Internet servers, rather than on researchers’ local http://wendang.chazidian.comStorageputerFirewallCloulad ptformLdaargtae ??lesDATA HIGHWAY Once a genome sequencer has cranked out its snippets of genomic information, or ‘reads’, they must be assembled into a continuous stretch of DNA using computing and software. Xu and his team try to automate as much of this process as possible to enable scientists to get to analyses quickly. Next, either the reads or the analysis, or both, have to travel to scientists. Generally, researchers share biological data with their peers through public repositories, such as the EBI or ones run by the US National Center for Biotechnology Information in Bethesda, Mary land. Given the size of the data, this travel often means physically delivering hard drives — and risks data getting lost, stolen or damaged. Instead, BGI wants to use either its own clouds or others of the customer’s choos-ing for electronic delivery. But that presents a problem, because big-data travel often means big traffic jams. Currently, BGI can transfer about 1 tera-byte per day to its customers. “If you transfer one genome at a time, it’s OK,” says Xu. “If you sequence 50, it’s not so practical for us to trans-fer that through the Internet. That takes about 20 days.” BGI is exploring a variety of technologies to accelerate electronic data transfer, among them fasp, software developed by Aspera in Emeryville, California, which helps to deliver 258 | NATURE | VOL 498 | 13 JUNE 2013

内容需要下载文档才能查看

data for film-production studios and the oil and gas industry as well as the life sciences. In an experiment last year, BGI tested a fasp-enabled data transfer between China and the University of California, San Diego (UCSD). It took 30 seconds to move a 24-gigabyte file. “That’s really fast,” says Xu. Data transfer with fasp is hundreds of times quicker than methods using the normal Inter-net protocol, says software engineer Michelle Munson, chief executive and co-founder of Aspera. However, all transfer protocols share challenges associated with transferring large, unstructured data sets.The test transfer between BGI and UCSD was encouraging because Internet connec-tions between China and the United States are “riddled with challenges” such as variations in signal strength that interrupt data transfer, says Munson. The protocol has to handle such road bumps and ensure speedy transfer, data integrity and privacy. Data transfer often slows when the passage is “There’s no bumpy, but with fasp reason to move it does not. Trans-data outside the fers can fail when a file is partially sent; cloud. You can do analysis right with ordinary Inter-net connections, there.”this relaunches the entire transfer. By contrast, fasp restarts where the previous transfer stopped. Data that are already on their way do not get resent, but continue on their travels. Xu says that he liked the experiment with fasp, but the software does not solve the data-transfer problem. “The main problem is not technical, it is economical,” he says. BGI would need to maintain a large Internet con-nection bandwidth for data transfer, which would be prohibitively expensive, especially given that Xu and his team do not send out big data in a continuous flow. “If we only transfer

© 2013 Macmillan Publishers Limited. All rights reservedperiodically, it doesn’t make any economic sense for us to have this infrastructure, espe-cially if the user wants that for free,” he says. Data-sharing among many collaborators also remains a challenge. When BGI uses fasp to share data with customers or collaborators, it must have a software licence, which allows customers to download or upload the data for free. But customers who want to share data with each other using this transfer protocol will need their own software licences. Putting the data on the cloud and not moving them would bypass this problem; teams would go to the large data sets, rather than the other way around. Xu and his team are exploring this approach, alongside the use of Globus Online, a free Web-based file-transfer service from the Computation Institute at the University of Chicago and the Argonne National Laboratory in Illinois. In April, the Computation Institute team launched a genome-sequencing-analysis service called Globus Genomics on the Amazon cloud.Munson says that Aspera has set up a pay-as-you-go system on the Amazon cloud to address the issue of data-sharing. Later this year, the company will begin selling an updated version of its software that can be embedded on the desktop of any kind of com-puter and will let users browse large data sets much like a file-sharing application. Files can be dragged and dropped from one location to another, even if those locations are commercial or academic clouds. The cost of producing, acquiring and dis-seminating data is decreasing, says James Taylor, a computational biologist at Emory University in Atlanta, Georgia, who thinks that “everyone should have access to the skills and tools” needed to make sense of all the information. Taylor is a co-founder of an academic platform called Galaxy, which lets scientists analyse their data and share soft-ware tools and workflows for free. Through

Web-based access to computing facilities at Pennsylvania State University (PSU) in Uni-versity Park, scientists can download Galaxy’s platform of tools to their local hardware, or use it on the Galaxy cloud. They can then plug in their own data, perform analyses and save the steps in them, or try out workflows set up by their colleagues. Spearheaded by Taylor and Anton Nekrutenko, a molecular biologist at PSU, the Galaxy project draws on a community of around 100 software developers. One feature is Tool Shed, a virtual area with more than 2,700 software tools that users can upload, try out and rate. Xu says that he likes the col-lection and its ratings, because without them, scientists must always check if a software tool actually runs before they can use it. valuable feedback to their computationally flu-ent colleagues because of different needs and approaches to the science, she says. Increasingly, big genomic data sets are being used in biotechnology companies, drug firms and medical centres, which also have specific needs. Robert Mulroy, president of Merrimack Pharmaceuticals in Cambridge, Massachu-setts, says that his teams handle mountains of data that hide drug candidates. “Our view is that biology functions through systems dynamics,” he says. Merrimack researchers focus on interro-gating molecular signalling networks in the healthy body and in tumours, hoping to find new ways to corner cancer cells. They generate and use large amounts of information from the genome and other factors that drive a cell to become cancerous, says Mulroy. The company KNOWLEDGE IS POWERstores its data and conducts analysis on its own Galaxy is a good fit for scientists with some computing infrastructure, rather than a cloud, computing know-how, says Alla Lapidus, a to keep the data private and protected. computational biologist in the algorithmic Drug developers have been hesitant about biology lab at St Petersburg Academic Univer-cloud computing. But, says Sundquist, that fear sity of the Russian Academy of Sciences, which is subsiding in some quarters: some companies is led by Pavel Pevzner, a computer scientist at that have previously avoided clouds because of UCSD. But, she says, the platform might not security problems are now exploring them. To be the best choice for less tech-savvy research-assuage these users’ concerns, Sundquist has ers. When Lapidus wanted to disseminate the engineered the DNAnexus cloud to be compli-software tools that she developed, she chose ant with US and European regulatory guide-to put them on DNAnexus’s newly launched lines. Its security features include encryption second-generation commercial cloud-based for biomedical information, and logs to allow analysis http://wendang.chazidian.comers to address potential queries from audi-That platform is also designed to cater to tors such as regulatory agencies, all of which is non-specialist users, says Sundquist. It is pos-important in drug development.sible for a computer scientist to build his or her own biological data-analysis suite with software CHALLENGES AND OPPORTUNITIEStools on the Amazon cloud, but DNAnexus Harnessing powerful computers and numer-uses its own engineering to help researchers ous tools for data analysis is crucial in drug dis-without the necessary computer skills to get to covery and other areas of big-data biology. But the analysis steps.that is only part of the problem. Data and tools Catering for non-specialists is important need to be more than close — they must talk to when developing tools, as well as platforms. The one another. Lapidus says that results produced Biomedical Information Science and Technol-by one tool are not always in a format that can ogy Initiative (BISTI) run by the US National Institutes of Health (NIH) in Bethesda, Mary-land, supports development of new computa-tional tools and the maintenance of existing ones. “We want a deployable tool,” says Vivien Bonazzi, programme director in computational biology and bioinformatics at the National Human Genome Research Institute, who is involved with BISTI. Scientists who are not heavy-duty informatics types need to be able to set up these tools and use them successfully, she says. And it must be possible to scale up tools and update them as data volume grows.Bonazzi says that although many life scientists have significant computational skills, others do not understand computer lingo enough to know that in the tech world, Python is not a snake and Perl is not a gem (they are programming languages). But even if biologists can’t develop or adapt the software, says Bonazzi, they have a place in big-data sci-Various data-transfer protocols handle problems ence. Apart from anything else, they can offer in different ways, says Michelle Munson.

© 2013 Macmillan Publishers Limited. All rights reservedSTANFORD UNIV./DNANEXU

内容需要下载文档才能查看

SArend Sidow wants to move data mountains without feeling pinched by infrastructure.be used by the next tool in a workflow. And if software tools are not easily installed, computer specialists will have to intervene on behalf of those biologists without computer skills. Even computationally savvy researchers can get tangled up when wrestling with software and big data. “Many of us are getting so busy analysing huge data sets that we don’t have time to do much else,” says Steven Salzberg, a computational biologist at Johns Hopkins University in Baltimore, Maryland. “We have to spend some of our time figuring out ways to make the analysis faster, rather than just using the tools we have.” Yet other big-data pressures come from the need to engineer tools for stability and longevity. Too many software tools crash too often. “Everyone in the field runs into similar problems,” says Hunter. In addition, research teams may not be able to acquire the resources they need, he says, especially in countries such as the United States, where an academic does not gain as much recognition for soft-ware engineering as for publishing a paper. With its dedicated focus on data and software infrastructure designed to serve scientists, the EBI offers an “interesting contrast to the US model”, says Hunter. US funding agencies are not entirely ignor-ing software engineering, however. In addi-tion to BISTI, the NIH is developing Big Data to Knowledge (BD2K), an initiative focused on managing large data sets in biomedicine, with elements such as data handling and standards, informatics training and software sharing. And as the cloud emerges as a popular place to do research, the agency is also reviewing data-use policies. An approved study usually lays out specific data uses, which may not include placing genomic data on a cloud, says Bonazzi. When a person consents to have his or her data used in one way, researchers cannot suddenly change that use, she says. In a big-data age that uses the cloud in addition to local hardware,

内容需要下载文档才能查看

new technologies in encryption and secure 13 JUNE 2013 | VOL 498 | NATURE | 259

内容需要下载文档才能查看

ASPERA

transmission will need to address such privacy concerns. Big data takes large numbers of people. BGI employs more than 600 engineers and software developers to manage its information-technol-ogy infrastructure, handle data and develop software tools and workflows. Scores of infor-maticians look for biologically relevant mes-sages in the data, usually tailored to requests from researchers and commercial customers, says Xu. And apart from its stream of research collaborations, BGI offers a sequencing and analysis service to customers. Early last year, the institute expanded its offerings with a cloud-based genome-analysis platform called Easy Genomics. In late 2012, it also bought the faltering US company Complete Genomics (CG), which offered human genome sequencing and analysis for customers in academia or drug discovery. Although the sale dashed hopes for earnings among CG’s investors, it doesn’t seem to have dimmed their view of the pros-pects for sequencing and analysis services. “It is now just a matter of time before sequencing data are used with regularity in clinical prac-tice,” says one investor, who did not wish to be identified. But the sale shows how difficult it can be to transition ideas into a competitive marketplace, the investor says. When tackling data mountains, BGI uses not only its own data-analysis tools, but also some developed in the academic community. To ramp up analysis speed and capacity as data sets grow, BGI assembled a cloud-based series of analysis steps into a workflow called Gaea, which uses the Hadoop open-source software framework. Hadoop was written by volunteer developers from companies and universities, and can be deployed on various types of com-puting infrastructure. BGI programmers built on this framework to instruct software tools to perform large-scale data analysis across “The cultural many computers at baggage of the same time. biology that If 50 genomes are privileges data to be analysed and generation the results com-over all other pared, hundreds of forms of science computational steps is holding us

内容需要下载文档才能查看

are involved. The back.”steps can run either sequentially or in parallel; with Gaea, they run in parallel across hundreds of cloud-based computers, reducing analysis time rather like many people working on a single large puzzle at once. The data are on the BGI cloud, as are the tools. “If you perform analysis in a non-parallel way, you will maybe need two weeks to fully process those data,” says Xu. Gaea takes around 15 hours for the same number of data. To leverage Hadoop’s muscle, Xu and his team needed to rewrite software tools. But the investment is worth it because the Hadoop framework allows analysis to continue as the 260 | NATURE | VOL 498 | 13 JUNE 2013

内容需要下载文档才能查看

MERRIMACK PHARMACEUTICALSA simplified array of breast-cancer subtypes, produced by researchers at Merrimack Pharmaceuticals, who use their own computational infrastructure to hunt for new cancer drugs. data mountains grow, he says. They are still ironing out some issues with Gaea, comparing its performance on the cloud with its performance on local infrastructure. Once testing is complete, BGI plans to mount Gaea on a cloud such as Amazon for use by the wider scientific community. Other groups are also trying to speed up analysis to cater to scientists who want to use big data. For example, Bina Technologies in Redwood City, California, a spin-out from Stan-ford University and the University of Califor-nia, Berkeley, has developed high-performance computing components for its genome-analysis services. Customers can buy the hardware, called the Bina Box, with software, or use Bina’s analysis platform on the cloud. viruses in the genomes of other species, includ-ing humans. Her work generates terabytes of data, which she shares with other researchers. Given that big-data analysis in biology is incredibly difficult, Hunter says, open science is becoming increasingly important. As he explains, researchers need to make their data available to the scientific community in a use-ful form, for others to mine. New science can emerge from the analysis of existing data sets: McClure generates some of her findings from other people’s data. But not everyone recog-nizes that kind of biology as an equal. “The cultural baggage of biology that privileges data generation over all other forms of science is holding us back,” says Hunter. A number of McClure’s graduate students are microbial ecologists, and she teaches them FROM VIVO TO SILICO how to rethink their findings in the face of so Data mountains and analysis are altering the many new data. “Before taking my class, none way science progresses, and breeding biologists of these students would have imagined that who get neither their feet nor their hands wet. they could produce new, meaningful knowl-“I am one of a small original group who made edge, and new hypotheses, from existing data, the first leap from the wet world to the in silico not their own,” she says. Big data in biology world to do biology,” says Marcie McClure, a add to the possibilities for scientists, she says, computational biologist at Montana State Uni-because data sit “under-analysed in databases versity in Bozeman. “I never looked back,” all over the world”. ■During her graduate training, McClure ana-lysed a class of viruses known as retroviruses in Vivien Marx is technology editor at Nature fish, doing the work of a “wet-worlder”, as she and Nature Methods.calls it. Since then, she and her team have dis-covered 11 fish retroviruses without touching 1. Mattmann, C. Nature 493, 473–475 (2013).2. Greene, C. S. & Troyanskaya, O. G. PLoS Comput. water in lake or lab, by analysing genomes com-Biol. 8, e1002816 (2012).putationally and in ways that others had not. She 3. EMBL–European Bioinformatics Institute EMBL-EBI has also developed software tools to find such Annual Scientific Report 2012 (EMBL–EBI, 2013).

© 2013 Macmillan Publishers Limited. All rights reserved

版权声明:此文档由查字典文档网用户提供,如用于商业用途请与作者联系,查字典文档网保持最终解释权!

下载文档

热门试卷

2016年四川省内江市中考化学试卷
广西钦州市高新区2017届高三11月月考政治试卷
浙江省湖州市2016-2017学年高一上学期期中考试政治试卷
浙江省湖州市2016-2017学年高二上学期期中考试政治试卷
辽宁省铁岭市协作体2017届高三上学期第三次联考政治试卷
广西钦州市钦州港区2016-2017学年高二11月月考政治试卷
广西钦州市钦州港区2017届高三11月月考政治试卷
广西钦州市钦州港区2016-2017学年高一11月月考政治试卷
广西钦州市高新区2016-2017学年高二11月月考政治试卷
广西钦州市高新区2016-2017学年高一11月月考政治试卷
山东省滨州市三校2017届第一学期阶段测试初三英语试题
四川省成都七中2017届高三一诊模拟考试文科综合试卷
2017届普通高等学校招生全国统一考试模拟试题(附答案)
重庆市永川中学高2017级上期12月月考语文试题
江西宜春三中2017届高三第一学期第二次月考文科综合试题
内蒙古赤峰二中2017届高三上学期第三次月考英语试题
2017年六年级(上)数学期末考试卷
2017人教版小学英语三年级上期末笔试题
江苏省常州西藏民族中学2016-2017学年九年级思想品德第一学期第二次阶段测试试卷
重庆市九龙坡区七校2016-2017学年上期八年级素质测查(二)语文学科试题卷
江苏省无锡市钱桥中学2016年12月八年级语文阶段性测试卷
江苏省无锡市钱桥中学2016-2017学年七年级英语12月阶段检测试卷
山东省邹城市第八中学2016-2017学年八年级12月物理第4章试题(无答案)
【人教版】河北省2015-2016学年度九年级上期末语文试题卷(附答案)
四川省简阳市阳安中学2016年12月高二月考英语试卷
四川省成都龙泉中学高三上学期2016年12月月考试题文科综合能力测试
安徽省滁州中学2016—2017学年度第一学期12月月考​高三英语试卷
山东省武城县第二中学2016.12高一年级上学期第二次月考历史试题(必修一第四、五单元)
福建省四地六校联考2016-2017学年上学期第三次月考高三化学试卷
甘肃省武威第二十三中学2016—2017学年度八年级第一学期12月月考生物试卷

网友关注视频

19 爱护鸟类_第一课时(二等奖)(桂美版二年级下册)_T502436
三年级英语单词记忆下册(沪教版)第一二单元复习
【部编】人教版语文七年级下册《泊秦淮》优质课教学视频+PPT课件+教案,广东省
沪教版八年级下册数学练习册21.3(2)分式方程P15
【部编】人教版语文七年级下册《老山界》优质课教学视频+PPT课件+教案,安徽省
北师大版数学 四年级下册 第三单元 第二节 小数点搬家
沪教版牛津小学英语(深圳用) 四年级下册 Unit 2
第19课 我喜欢的鸟_第一课时(二等奖)(人美杨永善版二年级下册)_T644386
二次函数求实际问题中的最值_第一课时(特等奖)(冀教版九年级下册)_T144339
青岛版教材五年级下册第四单元(走进军营——方向与位置)用数对确定位置(一等奖)
第五单元 民族艺术的瑰宝_16. 形形色色的民族乐器_第一课时(岭南版六年级上册)_T3751175
外研版英语七年级下册module1unit3名词性物主代词讲解
苏科版数学八年级下册9.2《中心对称和中心对称图形》
六年级英语下册上海牛津版教材讲解 U1单词
飞翔英语—冀教版(三起)英语三年级下册Lesson 2 Cats and Dogs
北师大版小学数学四年级下册第15课小数乘小数一
【部编】人教版语文七年级下册《过松源晨炊漆公店(其五)》优质课教学视频+PPT课件+教案,辽宁省
化学九年级下册全册同步 人教版 第25集 生活中常见的盐(二)
冀教版英语四年级下册第二课
精品·同步课程 历史 八年级 上册 第15集 近代科学技术与思想文化
二年级下册数学第二课
外研版八年级英语下学期 Module3
【部编】人教版语文七年级下册《老山界》优质课教学视频+PPT课件+教案,安徽省
冀教版小学数学二年级下册第二单元《租船问题》
苏科版八年级数学下册7.2《统计图的选用》
【部编】人教版语文七年级下册《泊秦淮》优质课教学视频+PPT课件+教案,天津市
北师大版数学四年级下册3.4包装
冀教版小学英语四年级下册Lesson2授课视频
8 随形想象_第一课时(二等奖)(沪教版二年级上册)_T3786594
沪教版牛津小学英语(深圳用) 五年级下册 Unit 7