Home | About | Sematext search-lucene.com search-hadoop.com
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB
 Search Hadoop and all its subprojects:

Switch to Threaded View
HBase >> mail # user >> save video with hbase


Copy link to this message
-
Re: Re: save video with hbase
I would save them in different rows. You could do that with a row key
that looks something like this:

<video name>\0<block number>

You can use any separator there, not just \0, that isn't a valid video
name or block number. This way you could skip to the block of the
video that you need, but you're not forcing all of the blocks of a
single video to be on one region server. Since regions are split on
row boundaries, you run into some of the same issues with having a
100M+ row as having 100M in a single column.

-Joey

2011/10/31 xtliwen <[EMAIL PROTECTED]>:
> You are right,100M is too big ,so I plan to split the video to multi parts(avg 5M ), and save in different colums.
>
>
> 2011-11-01
>
>
>
> xtliwen
>
>
>
> 发件人: Doug Meil
> 发送时间: 2011-11-01  09:54:47
> 收件人: [EMAIL PROTECTED]
> 抄送:
> 主题: Re: save video with hbase
>
> Hi there, see...
> http://hbase.apache.org/book.html#supported.datatypes
> ... while you can store anything that can be converted to a byte-array in
> HBase, as JD says 100meg is too big.
> On 10/31/11 9:09 PM, "xtliwen" <[EMAIL PROTECTED]> wrote:
>>
>>I would like to save the data using hbase.The hbase is designed for small
>>data,such as log.Maybe the video is too large for a
>>row,so I am worried about the performance of the hbase in storing the
>>video data.
>>
>>2011-11-01
>>
>>
>>
>>xtliwen
>>
>>
>>
>>发件人: Jean-Daniel Cryans
>>发送时间: 2011-11-01  08:07:02
>>收件人: user
>>抄送:
>>主题: Re: save video with hbase
>>
>>Why would you use HBase for that?
>>Regarding your two questions:
>>On Mon, Oct 31, 2011 at 4:17 PM, xtliwen <[EMAIL PROTECTED]> wrote:
>>> Hi everybody,
>>> When the client visit the video of a website through my website, it
>>>will be transcoded with our video codec server.As the time goes on,the
>>>quantities of the videos is quite large. Usuallly, a video can be
>>>tanscoded to serveral levels,so a original video is corresponding to
>>>multi transcoded videos . Now,we plan to record the video files with
>>>hbase. There are two problems with the hbase:
>>> 1 the video file is too large (100M avg.)
>>Why do you say that? The client side limit by default is -1
>>(disabled), and if there's any other limit they are just set
>>arbitrarily.
>>> 2 we require the transcoded video must can be read while it was been
>>>writing
>>That sounds like an application-level problem, not HBase.
>>>
>>> So,can anybody give some suggestions, Thanks.
>>>
>>> 2011-10-31
>>>
>>>
>>>
>>> regards
>>> xtliwen
>>>
>

--
Joseph Echeverria
Cloudera, Inc.
443.305.9434
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB