TW201301199A - Edge-based video interpolation for video and image upsampling - Google Patents
Edge-based video interpolation for video and image upsampling Download PDFInfo
- Publication number
- TW201301199A TW201301199A TW101104396A TW101104396A TW201301199A TW 201301199 A TW201301199 A TW 201301199A TW 101104396 A TW101104396 A TW 101104396A TW 101104396 A TW101104396 A TW 101104396A TW 201301199 A TW201301199 A TW 201301199A
- Authority
- TW
- Taiwan
- Prior art keywords
- pixel
- interpolation
- edge
- filter
- sharpening
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/403—Edge-driven scaling; Edge-based scaling
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Image Processing (AREA)
- Facsimile Image Signal Circuits (AREA)
Abstract
Description
相關申請案的交叉引用
本申請案要求(i)2011年2月11日提出的、名稱為“Edge-Based Video Interpolation For Video and Image Upsampling”的美國臨時專利申請案序號61/442,069和(ii)2011年9月15日提出的、名稱為“Combined Image Interpolation and Sampling”的美國臨時專利申請案序號61/535,353的優先權,這兩個專利申請案的內容都經由引用合併到本申請案中。
CROSS-REFERENCE TO RELATED APPLICATIONS RELATED APPLICATIONS [ s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s s The priority of U.S. Provisional Patent Application Serial No. 61/535,353, filed on Sep. 15, 2011, which is incorporated herein by reference in its entirety, the entire entire entire entire entire entire entire entire entire entire entire entire entire entire entire entire entire entire entire entire entire entire entire entire all
在影像和視訊處理中,放大或縮放(zoom)指的是增加數位影像或視訊的解析度的過程。通常,在大部分的視訊播放器(例如VLC播放器、Windows媒體播放器等)中都能夠發現放大。線上視訊主機網站(例如YouTube、Hulu、DailyMotion等)向用戶提供將視訊放大成全螢幕解析度的選項。許多目前的TV和DVD播放器都開始配備內建放大模組。
放大通常包括從現有的像素產生新的影像像素。在視訊訊框中,像素通常與其鄰近像素相干。此特性被大部分的放大方法用於產生新的像素。在視訊或影像轉碼器管線(pipeline)中,放大是有挑戰的,因為輸入是具有編碼偽像(coding artifact)的低解析度視訊。
放大的最簡單方法是像素複製,其簡單地涉及在列和行中複製像素。然而,此方法會導致嚴重的塊狀圖案(blockiness)。能夠藉由將低通濾波器應用到放大後的訊框中來減輕在像素複製中看到的塊效應和混疊偽像。用於降低塊效應的一些技術包括應用低通濾波器。通常被稱為內插的其他技術可以包括對鄰近像素進行平均以確定新的像素值。一種內插(interpolation)技術利用其後跟隨有低通濾波的“零填充”。這可以被描述為,藉由在中間位置處放置零值的像素來放大影像的尺寸,然後向放大後的影像應用低通濾波器。
另外,通常在視訊主機網站上找到的視訊有時被重壓縮成低位元率,因此可能具有低解析度。這些視訊中的常見偽像是模糊強度(blurriness)。放大過程通常增強了視訊中的模糊強度的能見度。因此,可以使用銳化濾波器來將一些細節恢復到視訊中。
傳統地,銳化包括對模糊影像進行高通濾波並將加權後的高通影像添加到原始影像中。這被稱為未銳化遮罩(unsharp masking)。高通濾波器有若干個實現方式,包括拉普拉斯(Laplacian)濾波器和拉普拉斯-高斯濾波器。In image and video processing, zooming in or zooming refers to the process of increasing the resolution of a digital image or video. In general, amplification can be found in most video players (such as VLC players, Windows Media Players, etc.). Online video hosting sites (such as YouTube, Hulu, DailyMotion, etc.) offer users the option to zoom in to full screen resolution. Many current TV and DVD players are beginning to be equipped with built-in amplification modules.
Amplification typically involves generating new image pixels from existing pixels. In the video frame, pixels are usually coherent with their neighboring pixels. This feature is used by most amplification methods to generate new pixels. In video or video transcoder pipelines, amplification is challenging because the input is low resolution video with coding artifacts.
The easiest way to zoom in is pixel copying, which simply involves copying pixels in columns and rows. However, this method can lead to severe blockiness. Blockiness and aliasing artifacts seen in pixel duplication can be mitigated by applying a low pass filter to the magnified frame. Some techniques for reducing blockiness include the application of low pass filters. Other techniques, commonly referred to as interpolation, may include averaging adjacent pixels to determine new pixel values. An interpolation technique utilizes "zero fill" followed by low pass filtering. This can be described as amplifying the size of the image by placing a zero-valued pixel at an intermediate position, and then applying a low-pass filter to the magnified image.
In addition, video that is typically found on video host sites is sometimes recompressed to a low bit rate and may therefore have low resolution. A common artifact in these videos is blurriness. The magnification process generally enhances the visibility of the blur intensity in the video. Therefore, a sharpening filter can be used to restore some details to the video.
Traditionally, sharpening involves high pass filtering the blurred image and adding the weighted high pass image to the original image. This is called unsharp masking. The high pass filter has several implementations, including a Laplacian filter and a Laplacian-Gaussian filter.
本文描述的一個實施方式是用於對影像進行內插的方法,該方法包括:確定與內插點相關聯的邊緣特性,該邊緣特性具有邊緣大小和邊緣角;回應於邊緣角來選擇內插濾波器;以及使用所選擇的內插濾波器來確定內插點處的像素值。邊緣特性可以基於確定與內插點相關聯的鄰近區域中的像素值的水平梯度和垂直梯度。與內插點相關聯的鄰近區域可以是水平矩形區域、垂直矩形區域或正方形區域。在一個實施方式中,回應於內插點為列內插點、行內插點或中心內插點來確定鄰近區域。在一些實施方式中,通過使用第一階梯度濾波器(其可以被稱為“蒙板(mask)”,因為沒有執行卷積操作)來確定邊緣特性。一種這樣的第一階梯度濾波器是修正的Sobel(索貝爾)運算子。
該方法還可以包括回應於邊緣特性來選擇內插濾波器。可以回應於邊緣大小而進行內插濾波器的選擇。也就是說,邊緣大小(或估計的大小)大於臨界值的影像點可以使用一組內插濾波器(例如,被實現為運行於微處理器上的硬體電路或軟體或其組合),而邊緣大小(或估計的大小)低於臨界值的影像點則可以使用另一組內插濾波器。在一些實施方式中,也可以依據影像的特性及/或邊緣特性來調整該臨界值。
該方法可以利用這樣的內插濾波器,其將較大的權重應用到位於沿著邊緣角的方向上的像素上、並將較小的權重應用到位於沿著與邊緣角垂直的方向上的像素上或者不向位於沿著與邊緣角垂直的方向上的像素應用權重。在一個實施方式中,內插濾波器對最近的鄰近像素應用較大的權重,對位於沿著邊緣角的方向上的像素應用中等權重,以及對位於沿著與邊緣角垂直的方向上的像素應用最小的權重或不應用權重。在一個實施方式中,最近的鄰近像素是與內插點位於同一列或同一行中的像素。
這裏描述的該方法可以在處理器上實施。這樣,電腦可讀媒體可以用於儲存指令,當這些指令被處理器執行時電腦可讀媒體將促使處理器:獲取多個邊緣特性,每個邊緣特性與多個內插點中的各自的一個內插點相關聯,並且每個邊緣特性具有邊緣大小和邊緣角;回應於各自的邊緣角,為多個內插點中的每個內插點選擇內插濾波器;使用相應的被選擇的內插濾波器來確定針對多個內插點中的每個內插點的像素值;以及輸出該像素值。
在一些實施方式中,該方法被實施於專用硬體中、或者使用硬體來加速計算的子集。在一個實施方式中,內插裝置包括:邊緣特性計算器,被配置為確定多個內插點中的每個內插點的邊緣特性;內插濾波器選擇器,被配置為對邊緣特性進行運算並回應性地產生用於多個內插點中的每個內插點的內插濾波器識別符;以及內插濾波器電路,被配置為回應於內插濾波器識別符來應用多個內插濾波器中的一個內插濾波器、並輸出針對多個內插點的被內插的值。這些元件中的一個或多個元件可以藉由用於執行儲存在電腦可讀儲存媒體中的代碼的處理器來實施。
在再一實施方式中,描述了其後跟隨有自適應銳化濾波器的用於放大的基於邊緣的內插。在一個實施方式中,銳化濾波器由基於邊緣的內插參數控制,該參數確定要被銳化的像素和銳化強度。在一個實施方式中,該方法包括:確定要被內插的影像像素的梯度資料;根據梯度資料,使用鄰近像素來確定被內插的像素值;基於梯度資料來識別要被銳化的影像像素;以及使用部分基於梯度資料所確定的濾波器強度來銳化影像像素。
在替代的實施方式中,基於邊緣的內插濾波器操作與自適應銳化濾波器組合成稱為聯合濾波器的單一濾波器。在一個實施方式中,該方法包括:確定要被內插的影像像素的梯度資料;選擇性地銳化鄰近的原始像素;回應於是中心列或行像素的像素類別(category)來選擇性地識別還要被內插的鄰近像素;以及根據聯合銳化和內插濾波器,使用鄰近像素來確定內插和銳化後的像素值。One embodiment described herein is a method for interpolating an image, the method comprising: determining an edge characteristic associated with the interpolated point, the edge characteristic having an edge size and an edge angle; selecting the interpolation in response to the edge angle a filter; and using the selected interpolation filter to determine a pixel value at the interpolation point. The edge characteristics may be based on determining a horizontal gradient and a vertical gradient of pixel values in adjacent regions associated with the interpolation points. The neighboring area associated with the interpolated point may be a horizontal rectangular area, a vertical rectangular area, or a square area. In one embodiment, the neighboring region is determined in response to the interpolated point being a column interpolated point, an interpolated point, or a central interpolated point. In some embodiments, the edge characteristics are determined by using a first step filter (which may be referred to as a "mask" because the convolution operation is not performed). One such first step filter is a modified Sobel operator.
The method can also include selecting an interpolation filter in response to the edge characteristic. The selection of the interpolation filter can be made in response to the edge size. That is, an image point whose edge size (or estimated size) is greater than a threshold may use a set of interpolation filters (eg, implemented as a hardware circuit or software running on a microprocessor or a combination thereof). Another set of interpolation filters can be used for image points whose edge size (or estimated size) is below the threshold. In some embodiments, the threshold may also be adjusted depending on the characteristics of the image and/or the edge characteristics.
The method may utilize an interpolation filter that applies a larger weight to pixels located in a direction along the edge angle and applies a smaller weight to a direction that is perpendicular to the edge angle Weights are applied to pixels that are located in a direction that is perpendicular to the edge angle. In one embodiment, the interpolation filter applies a greater weight to the nearest neighboring pixels, a medium weight to the pixels located in the direction along the edge angle, and a pixel located in a direction perpendicular to the edge angle Apply the smallest weight or no weight. In one embodiment, the nearest neighboring pixels are pixels that are in the same column or in the same row as the interpolated points.
The method described herein can be implemented on a processor. Thus, a computer readable medium can be used to store instructions that, when executed by a processor, cause the processor to: acquire a plurality of edge characteristics, each edge characteristic and a respective one of a plurality of interpolation points The interpolation points are associated, and each edge characteristic has an edge size and an edge angle; in response to the respective edge angles, an interpolation filter is selected for each of the plurality of interpolation points; using the corresponding selected ones Interpolating a filter to determine a pixel value for each of the plurality of interpolated points; and outputting the pixel value.
In some embodiments, the method is implemented in dedicated hardware or uses hardware to speed up the subset of calculations. In one embodiment, the interpolation device includes an edge characteristic calculator configured to determine edge characteristics of each of the plurality of interpolation points, and an interpolation filter selector configured to perform edge characteristics Interpolating and responsively generating an interpolation filter identifier for each of the plurality of interpolation points; and an interpolation filter circuit configured to apply the plurality of responses in response to the interpolation filter identifier Interpolating one of the interpolation filters and outputting the interpolated values for the plurality of interpolation points. One or more of these elements can be implemented by a processor for executing code stored in a computer readable storage medium.
In yet another embodiment, edge-based interpolation for amplification followed by an adaptive sharpening filter is described. In one embodiment, the sharpening filter is controlled by edge-based interpolation parameters that determine the pixels to be sharpened and the sharpening strength. In one embodiment, the method includes: determining gradient data of image pixels to be interpolated; determining adjacent pixel values using neighboring pixels according to gradient data; and identifying image pixels to be sharpened based on gradient data And use some of the filter strength determined based on the gradient data to sharpen the image pixels.
In an alternative embodiment, the edge based interpolation filter operation is combined with an adaptive sharpening filter into a single filter called a joint filter. In one embodiment, the method includes: determining gradient data of image pixels to be interpolated; selectively sharpening adjacent original pixels; selectively identifying in response to a pixel category of a center column or row of pixels The neighboring pixels to be interpolated; and the neighboring pixels are used to determine the interpolated and sharpened pixel values according to the joint sharpening and interpolation filters.
本文描述的內插方法和裝置可以被用於有線或無線網路中。包括手持裝置、桌上型電腦、膝上型電腦或其他電腦在內的裝置可以用於執行該方法。這包括蜂巢式電話、PDA、平板電腦及/或顯示器以及有線電視機上盒、電視等。
這裏描述的內插或視訊放大方案保持邊緣的保真度和計算複雜度低於典型的FIR濾波技術。該系統可以實施被分別施加給視訊的寬度和高度的任意放大因數N和M。簡單起見,這裏描述的是藉由這兩個維度中的因數2(N=2和M=2)對影像或視訊訊框進行放大。
第1圖示出了包含原始像素(用正方形表示)和估計像素或在內插點處被估計的像素值(用圓形表示)的像素格點。像素A具有對角的原始像素;像素B具有沿著同一行的兩個鄰近的原始像素;以及像素C具有沿著同一列的兩個鄰近的原始像素。因此,可以將估計像素分類為3組:(a)具有對角的鄰近原始像素的像素(在第1圖中被標記為“A”),並且位於這些內插點處的估計像素可以被稱為中心像素;(b)將鄰近的原始行像素作為最近的鄰近像素的像素(在第1圖中標記為“B”),該估計像素可以被稱為行像素;以及(c)將鄰近的原始列像素作為最近的鄰近像素的像素(在第1圖中被標記為“C”),該估計像素可以被稱為列像素。
在一個實施方式中,該方法包括下述方面:邊緣偵測;邊緣角確定;以及像素估計。在另一實施方式中,該方法可以包括:確定與內插點相關聯的邊緣特性,其中該邊緣特性包括邊緣大小和邊緣角;回應於邊緣角來選擇內插濾波器;以及使用所選擇的內插濾波器來確定內插點處的像素值。
可將用於計算像素的水平和垂直梯度的任何軟體或硬體(總稱為運算子)用於邊緣偵測。在一個實施方式中,使用修正的Sobel運算子,因為其計算複雜度低。Sobel運算子包括用於計算計算水平和垂直梯度的兩個方形蒙板。這些梯度然後被用於計算或以其他方式獲得梯度的角度和邊緣的角度或其估計。標準Sobel運算子被應用到方形像素格點,而修正的Sobel運算子可以被應用到矩形像素格點。
當確定了邊緣特性時,估計像素分類或內插點分類可以被用於選擇不同的修正Sobel蒙板Gx和Gy。如第2A-2C圖所示,Gx和Gy是可以分別被用於與影像像素逐點相乘並對乘積求和以計算水平(Δx)和垂直(Δy)梯度(或其估計)的蒙板。
在一個實施方式中,梯度(或邊緣)的大小G藉由計算G = |Δx| + |Δy|來確定。如果G大於臨界值T邊緣,則邊緣可以存在。在G超過臨界值的情況中,邊緣梯度的角度可以如下述那樣確定。臨界值T邊緣的值可以取決於視訊的解析度、像素類別和其他因素。可以根據經驗來確定臨界值,並且已經被發現能夠執行地很好的一些臨界值在下面的表1中給出。臨界值的選擇可以不同:在一些實施方式中可以選擇固定臨界值,而在其他實施方式中,可以在逐訊框地基礎上或像素塊的基礎上調整臨界值。可以基於塊內的像素特性來自動計算自適應臨界值。在較低的解析度下,鄰近像素不太可能相干,從而導致較大的梯度。使用更小的臨界值將會使許多的像素被分類為邊緣。因此,為了降低不準確的分類,對於更小解析度的視訊可以使用更大的臨界值。
在各種實施方式中,可以回應於邊緣(或梯度)大小、邊緣(或梯度)角或這兩者來選擇內插濾波器。在一個實施方式中,可以為其中相應的邊緣大小G小於T邊緣的內插點選擇一個內插濾波器。在其他實施方式中,可以其中相應的邊緣具有等於零的Δx或Δy的內插點選擇一個內插濾波器。在每個情況中,不需要確定或提供邊緣角,並且用於在內插點處插入新的像素的內插濾波器如下所示:
a. 對於中心像素:(a0 + a1 + b0 + b1)/4
b. 對於行像素:(a1+b1)/2
c. 對於列像素:(b0+b1)/2
其中,a0、a1、b0和b1對應於第2A-2C圖中所標記的像素。注意,Δy=0和Δx=0分別對應於梯度角0和90度。
梯度角μ可以被計算為μ = tan-1(Δy/Δx)。由於邊緣角垂直於梯度角,所以邊緣角為Θ = μ + 90o。注意,由於可以藉由簡單的90度旋轉將邊緣和邊緣梯度相聯繫,所以如果考慮此關係的話,兩者可以互換使用。還應當注意,邊緣(或梯度)角的範圍可以降低為0-180度的範圍。
在一些實施方式中,可以回應於像素類別和邊緣角來確定在各自的內插點處使用的內插濾波器。在一個實施方式中,內插假設邊緣是線性的。對於彎曲邊緣,可以在像素估計期間檢查其他角。替代地,可以使用基於霍夫(Hough)變換的方法來偵測屬於曲線的邊緣像素,並且這些像素可以用於沿著彎曲邊緣來插入新的像素。
現在將描述針對內插點的像素估計過程,其中邊緣大小超過臨界值。在一個實施方式中,對於中心像素,基於邊緣角是否近似為45度來選擇內插濾波器。一個實施方式針對邊緣角位於45至135度的範圍使用一個中心內插濾波器。替代的實施方式可以利用其他的角範圍。
對於中心像素估計,可以如下述那樣執行內插:
如果θ介於35與55度之間,則新的像素=(a1+b0)/2。
如果θ介於125與145度之間,則新的像素=(a0+b1)/2。
其他情況,新的像素=(a0+a1+b0+b1)/4。
對於行像素估計,
如果θ介於0與90度之間,則新的像素=(b0 + 2*(a1+b1) + a2)/6。
其他情況 新的像素=(a0 + 2*(a1+b1) + b2)/6。
對於列像素估計,
如果θ介於0與90度之間,則新的像素=(c0 + 2 *(b0+b1) + a1)/6。
其他情況 新的像素=(a0 + 2 *(b0+b1) + c1)/6。
因此,該方法可以利用這樣的內插濾波器,其將較大的權重應用到位於沿著邊緣角的方向上的像素上、並將較小的權重應用到位於沿著與邊緣角垂直的方向上的像素上或者不向其應用權重。在一個實施方式中,內插濾波器應用較大的權重到最近的鄰近像素、應用中等權重到位於沿著邊緣角的方向上的像素、並應用最小的權重到位於沿著與邊緣角垂直的方向上的像素或不向其應用權重。在一個實施方式中,最近的鄰近像素是與內插點位於同一列或同一行中的像素。
在一個實施方式中,梯度測量Δx和Δy可以與查找表或LUT一起使用,以確定合適的內插濾波器。LUT可以儲存期望的濾波器脈衝回應、或者可以簡單地提供可以用於確定並應用合適的內插濾波器的內插濾波器識別符。
本文描述的一個實施方式是用於對影像進行內插的方法,該方法包括:確定與內插點相關聯的邊緣特性,該邊緣特性包括邊緣大小和邊緣角;回應於邊緣角來選擇內插濾波器;以及使用所選擇的內插濾波器來確定內插點處的像素值。邊緣特性可以基於確定與內插點相關聯的鄰近區域中的像素值的水平梯度和垂直梯度。與內插點相關聯的鄰近區域可以是如第2B圖所示的水平矩形區域、如第2C圖所示的垂直矩形區域或如第2A圖所示的方形區域。在一個實施方式中,回應於內插點為列內插點、行內插點或中心內插點來確定鄰近區域。在一些實施方式中,使用一階梯度濾波器(當不執行卷積操作時,其可以被稱為“蒙板(mask)”)來確定邊緣特性。一種這樣的一階梯度濾波器是修正的Sobel運算子。
下面描述基於邊緣的內插與使用表2中列出的視訊的五分接頭濾波器技術的性能比較。(在比較時,僅使用了五分接頭濾波器,但沒有使用銳化濾波器。)藉由將這兩種方法應用到代碼轉換後的視訊,來對這兩種方法進行測試。使用魯瑪(luma)分量的峰值信號雜訊比(PSNR)、視訊品質度量(VQM)評分和主觀比較來測量放大後的視訊的品質。轉碼器的輸入視訊被用作這裏的品質測量的參考。
接近於零的VQM評分表示沒有偽像/損傷,而接近於一的評分表示被嚴重損傷。VQM僅截取了15秒的視訊進行比較。對在不同的視訊和轉碼器位元率測試放大方案,並且在表3中列出了結果。表3中的項“2X”和“3X”指的是分別正在以原始位元率的一半和三分之一被編碼的代碼轉換後的視訊。對於“vid”視訊而言,這兩種放大方法都產生低的損傷,但是基於邊緣的內插稍微優於5分接頭濾波器。應當注意到,當觀看者兩個處理後的視訊時,基於邊緣的內插看起來明顯更好一些。這是預料的到的,因為大多數的客觀測量方案都具有局限性。與2X視訊相比,3X視訊常常產生稍微高的VQM評分,因為降低位元率會增加編碼失真。
對於艾利之書視訊而言,相較於5分接頭方法,基於邊緣的內插相比於5分接頭方法具有小的感知性改善。艾利之書的VQM評分高於“vid”視訊的VQM評分,因為以比原始視訊低50倍的位元率來編碼下採樣後的視訊。
對於“vid”視訊,針對不具有場景變化的視訊片段來計算PSNR,並在表4中報告結果。對於艾利之書,考慮PSNR≦ 50 dB的所有訊框,因為具有更高PSNR的所有訊框都有可能是空白訊框。對於vid15和vid16,相比於5分接頭濾波器,能夠獲得0.7 dB的最大PSNR改善。第3圖示出了針對vid01_2X序列的PSNR與訊框的快照(snapshot)。顯然,即使在逐訊框的基礎上,當使用基於邊緣的內插方法時,具有PSNR改善。
除了VQM評分,也藉由觀看放大後的視訊對這兩種放大方法進行比較。通過使視訊訊框經歷這兩種放大方案所獲得的實驗結果表明,與5分接頭濾波器方法相較之下,基於邊緣的內插方法具有更銳化的邊緣和更多的細節。
藉由在2.53 GHz PC上對1000訊框的每個測試序列運行放大演算法來進行複雜度比較,並測量每一訊框的平均編碼時間。藉由獲得5分接頭濾波器與基於邊緣的內插濾波器的處理時間的比率來計算相對速度改善。結果在表5中給出。我們發現,基於邊緣的內插比5分接頭濾波器要快最多2.5倍。
這裏還描述了用於內插和銳化視訊信號的裝置和方法。在一個實施方式中,該方法通過兩級濾波器方法來執行,其中內插濾波器將輸入視訊放大到更高的空間解析度,然後銳化濾波器增強視訊中的細節。在另一實施方式中,該方法藉由聯合濾波器(joint filter)方法來執行,其中一個濾波器被應用到輸入視訊信號以將該輸入視訊信號放大到更高的解析度並同時增強信號細節。聯合濾波器同時執行內插和銳化,並且可以具有相關聯的處理增益。在一些實施方式中,還可以降低記憶體存取。
如第4A圖所示,兩個分離的濾波器,即基於邊緣的內插濾波器和自適應銳化濾波器被組合以從原始視訊訊框源來產生放大和銳化後的視訊訊框。第4B圖中描述了一個實施方式,其中首先藉由使用基於邊緣的內插器在兩個維度中將YUV訊框放大兩倍。可以使用上面描述的基於邊緣的內插器。在訊框的亮度分量(Y)的內插期間所捕獲的邊緣資訊可以用於確定要被銳化的像素、和用於控制自適應銳化濾波器的強度。如第9圖中所概括的,可以從部分地在作為中心像素A、列像素B或行像素C的被內插的像素的位置選擇的邊緣偵測濾波器中獲得邊緣資訊。可以對魯瑪分量執行銳化,因為魯瑪分量通常包含更多的細節或邊緣。銳化後的魯瑪分量然後與色度分量相組合,以產生放大後的銳化訊框。還可以從色度分量中的一個或多個色度分量中擷取邊緣資訊,並且可以使用來自魯瑪和來自色度分量中的一個或多個色度分量的邊緣資訊的加權組合。另外,色度分量中的一個或多個色度分量也可以經歷自適應銳化。
第5圖和第6圖中的方塊圖分別示出了針對基於邊緣的內插和銳化階段的一個演算法。兩個緩衝器——即映射[N][M]和強度[N][M]——分別儲存銳化映射(map)和邊緣強度,其中N和M是放大後的訊框的寬度和高度。這兩個緩衝器在對訊框進行內插之前被初始化為零。映射[i][j]中的項“1”表明位於(i,j)處的像素要被銳化,否則為“0”。對於要被內插的每個像素(i,j),使用修正的Sobel運算子來計算梯度(gradij)。如果gradij≧銳化臨界值(sharp_thresh),則針對被內插的像素及其鄰近像素的映射[i][j]元素被設定為一(1),其中sharp_thresh是銳化臨界值。類似地,針對被內插的像素及其鄰近像素的強度[i][j]被設定為gradij。例如,如果位於(i, j)處的被內插的像素的梯度gradij≧sharp_thresh,則:
這在第8圖中示出,其中符號“x”表示原始像素,“o”表示被內插的像素。黑點是被標為用於銳化的被內插的邊緣像素,並且邊界框中的其他像素因其鄰近邊緣像素,所以也被標為用於銳化。在一個實施方式中,100的銳化臨界值是恰當的,但是臨界值的選擇可以不同。在其他實施方式中,銳化臨界值也可以從局部影像特性來確定、並且在整個影像中被調整為不同的值。
在獲得被內插的訊框以及強度[i][j]和映射[i][j]之後,將拉普拉斯-高斯(LoG)濾波器應用到(i,j)處映射[i][j]=1的像素。對邊緣像素的鄰近像素進行銳化,以移除因亮度的非均勻變化而可能看起來是斑點的偽像。在銳化階段中,再次使用來自內插階段的映射[i][j]和強度[i][j]資訊。這會導致較低的計算複雜度。
連續形式的LoG濾波器的定義如下所示:
其中,x和y是與LoG的中心(即(0,0))的距離,σ2是高斯方差。尺寸為KxL的離散LoG濾波器定義如下:
參數σ確定濾波強度。更大的σ(> 1)可以用於弱銳化,而更小的σ(≦ 1)可以用於更強的銳化。銳化濾波器還可以具有各種尺寸win_size。例如,對於具有小解析度(< 640x360)的輸入視訊,可以選擇win_size = 5x5的LoG濾波器;而對於更高解析度的視訊(≧ 640x360),可以選擇win_size = 9x9的LoG濾波器。濾波器尺寸確定了在銳化期間被考慮的鄰近像素的數量。在其他實施方式中,可以基於視訊解析度和內容而選擇不同的濾波器尺寸。另外,在整個影像訊框可以調整濾波器尺寸。作為示例,5x5的LoG濾波器的操作可以表示如下:
其中,C[i][j]和Ch[i][j]分別是(i,j)處的放大後的像素及其相應的高頻分量。第10圖示出了向其應用濾波器係數的像素的跨度,其中黑點表示要使用5x5的LoG濾波器進行銳化的被內插的像素(稱為中心像素)。LoG濾波器係數被應用到邊界框中的像素上。
在一個實施方式中,基於下述的邊緣強度來選擇σij:
其中,ρ1,…,ρS是邊緣強度臨界值,而α1, …,αS是與不同的銳化等級相關聯的S不同的σij值。在一個實施方式中,σij定義如下:
除了邊緣梯度,還可以使用局部影像特性來確定σij。例如,強邊緣的過度銳化(即,對強邊緣使用小的σij值)會導致振鈴(ringing)偽像。為了避免這種情況,當邊緣強度強度[i][j]增大時,可以使用更大的σij值。
最後,銳化後的像素Cout[i][j]如下獲得:
一個實施方式中使用λ=1,但是在其他實施方式中,可以使用不同的λ值。在其他實施方式中,也能夠藉由在內插階段之前偵測雜訊並使用此資訊來選擇具有更大σ(σ > 1)的LoG濾波器而使銳化濾波器對雜訊敏感。這將確保被偵測為邊緣的雜訊不被放大。類似地,在其他實施方式中,可以使銳化過程對偽邊緣(諸如塊偽像)敏感。可以在內插之前使用塊偽像偵測演算法。在銳化過程期間,被分類為偽邊緣的像素可以被平滑而非被銳化。
第7圖描繪了聯合的基於邊緣的內插和自適應銳化濾波器的實施方式。在之前的兩級濾波器方法中,在應用銳化濾波器之前在濾波器視窗中計算被內插的像素。在聯合濾波器方法中,內插與銳化或之間的依賴性降低了。在此實施方式中,基於邊緣的內插與銳化濾波器組合以提供聯合濾波器。第11圖中給出了組合後的濾波方法的流程圖。使用修正的Sobel運算子來計算針對位於(i,j)處的像素的梯度gradij。使用兩個緩衝器-即緩衝器銳化旗標(sharp_flag)和將來梯度(future_grad)來標出位於(k,n)處的尚未被內插的鄰近像素的銳化。如果梯度gradij大於或等於預定義的銳化臨界值sharp_thresh,則對鄰近的3x3像素執行下述的兩個操作,如第12圖所示:
1、使用LoG濾波器來銳化鄰近的原始像素,其中使用等式(5)以基於gradij來選擇LoG濾波器的σ。在第12圖中,這些像素由“x”指示。
2、將尚未被內插的鄰近像素的sharp_flag和future_grad設定如下:
如果像素(i,j)是中心像素,則設定:
。
如果像素(i,j)是列像素,則設定:
。
如果像素(i,j)是行像素,則設定:
。
在一個實施方式中,sharp_thresh = 100,並且像素銳化候選是鄰近的3x3像素。在其他實施方式中,考慮更大的鄰域,並且可以使用不同的sharp_thresh;例如,sharp_thresh的值可以基於局部影像特性來確定。如果gradij小於sharp_thresh,則檢查相關聯的sharp_flag[i][j]是否等於1。如果sharp_flag[i][j]等於1,則使用其相關聯的future_grad[i][j]值來選擇適當的聯合濾波器,否則繼續使用gradij來選擇聯合濾波器。藉由使用future_grad[i][j],該系統使用(i-a,j-b)處的之前被內插的鄰近像素的梯度,其中gradi-a j-b≧ sharp_thresh並且0≦a,b≦1。
如第12圖所描繪的,黑點是正在被內插和銳化的像素、並且具有大於銳化臨界值的梯度。“x”是要被銳化的原始像素,並且灰點是要被內插並且被標為用於銳化的像素。存在三種可能的情況:(a)中心、(b)列和(c)行像素內插和銳化。
下面是聯合濾波器的兩個實施方式。在其他實施方式中,除了原始像素,聯合濾波器還部分地對已經被內插的像素進行操作。
在上述的兩級濾波器方法中,win_size內的像素用於如第13圖中的框所示那樣銳化。在此實施方式中,如第13圖中的粗體X所表示的那樣,識別用於獲得邊界框中的被內插的像素的所有原始像素。銳化濾波器對這些原始像素進行操作。由於這些像素被用於獲取邊界框中的被內插的像素,所以從這些原始像素計算的高頻分量將接近於從邊界框中的所有像素所計算的高頻分量。
如下所示,聯合濾波器可以被寫為兩個濾波器的和:
其中,LOG是僅被應用到原始像素的銳化濾波器,並且hEI是基於邊緣的內插器。在此實施方式中,λ=1,但是在其他實施方式中可以選擇不同的λ值。LoG參數σ是邊緣強度的函數,諸如等式(5)中定義的函數。對於輸入視訊解析度小於640 x 360的一個實施方式,設定win_size = 5x5,否則將其設定為9x9。
基於邊緣的內插器hEI是梯度、邊緣角、濾波器尺寸和像素類別的函數。如上面參照第9圖描述的那樣,對於基於邊緣的內插,基於被內插的像素在像素格點中的位置將被內插的像素分類為中心像素、列像素和行像素。因此,包括原始像素在內,放大後的訊框具有四個像素類別。每個被內插的像素的類別(例如,中心、列或行)都具有三個濾波器,每個濾波器對應於三個邊緣方向中的其中一個方向。對於列和行像素,這三個邊緣方向是無邊緣、0與90度之間的邊緣角和90與180度之間的邊緣角。對於中心像素,這三個邊緣方向是無邊緣、35與55度之間的邊緣角和125與145度之間的邊緣角。在銳化期間,考慮用於獲取win_size內的被內插的像素的原始像素。例如,為了銳化第13圖中的中心像素,對4x4的原始像素執行濾波。因此,4x4的LOG濾波器被考慮,從而產生等式(7)中的4x4聯合濾波器。聯合濾波器的尺寸用濾波器尺寸(win_size)表示,因此對於第13圖,濾波器尺寸(win_size) = 濾波器尺寸(5x5) = 4x4。類似地,如第14、15和16圖所示,針對列、行和原始像素的濾波器尺寸(5x5)分別是5x4、4x5和5x5。
在第13圖中,示出了由黑點表示的中心像素的內插和銳化。win_size內的被內插的像素從被表示為粗體“x”的原始像素獲取。4x4的LoG濾波器被應用到這些原始像素。在第14圖中,示出了由黑點表示的列像素的內插和銳化。5x4的LoG濾波器被應用到由粗體“x”表示的原始像素。在第15圖中,示出了由黑點表示的行像素的內插和銳化。4x5的LoG濾波器被應用到由粗體“x”表示的原始像素。在第16圖中,描繪了被標記為“A”的原始像素的銳化。5x5的LoG濾波器被應用到由粗體“x”表示的原始像素。
下面是用於小螢幕視訊解析度(例如,480x204)的中心像素的聯合濾波器的實施方式。該程序與用於列、行和中心像素的程序類似。可以做出小的修改以實現針對更大視訊解析度(例如,1080p、720p)的聯合濾波器設計。對於小螢幕視訊解析度(例如使用win_size = 5x5),得到的聯合濾波器尺寸為濾波器尺寸(win_size) = 4x4。
下面是用於基於邊緣方向來產生中心像素的內插濾波器:
其中,邊緣臨界值確定被內插的像素是否是邊緣像素。
中心像素內插濾波器的實施方式如下給出:
在一個實施方式中,邊緣臨界值> sharp_thresh。尺寸為濾波器尺寸(win_size) = KxL的具有標準差σ的LoG濾波器如等式(3)中定義。
針對等式(5)中的四個σ值(其中α1=∞、α2=1、α3=1.2和α4=1.4)的LoG濾波器如下所示:
用於中心像素的聯合濾波器則如下所示:
類似地,
當僅應用內插並且不應用銳化時,使用濾波器h中心 , ∞。用於所有其他像素類別的聯合濾波器可以用類似的方式獲得。在一個實施方式中,使用等式(5)中列出的四個不同的σ值和兩個不同的win_size值。因此,用於被內插的像素的聯合濾波器的數量是:
原始像素僅需要銳化,因此在其濾波器設計期間,並不包括基於邊緣的內插。用於原始像素的聯合濾波器的數量是:
所需要的濾波器的總數量是:
這些濾波器可以儲存在查找表中,並且可以基於像素類別、梯度、邊緣角和win_size來選擇。
在替代的實施方式中,聯合濾波器與等式(7)相同,但是其使用不同的銳化分量。針對小螢幕視訊解析度(例如,480x204)的中心像素來描述替代的聯合濾波器的一個示例。相同的設計方法也可以擴展到列、行和原始像素。對於此示例,win_size = 5x5。第17圖示出了用於銳化的原始像素和所估計的被內插的像素,其中,win_size內的所有像素都用於銳化,並且被內插的像素(灰點)從原始像素(用粗體“x”表示)來估計。
在一個實施方式中,當梯度小於邊緣臨界值時,使用基於邊緣的內插來估計被內插的像素(第17圖中的灰點)。這對應於下面的針對中心、列和行像素的估計濾波器:
第17圖可以用下面的表6中的表格形式來表示,其中Cij是像素。灰色的陣列表示位於win_size = 5x5中的像素,並且粗體的項表示原始像素。因此,灰色區域中的非粗體像素是所估計的被內插的像素。
表6
5x5銳化濾波器表示如下。
可以根據下式來使用僅對原始像素進行操作但仍然能夠提供與等式(16)中相同的高頻分量Ch的銳化濾波器g:
為了說明g的設計,推導係數g00和g01。等式(16)中的每個乘積項用原始像素表示如下:
從等式(17),與C00相對應的係數的和是g00,並且與C02相對應的係數的和是g01。將等式(18)中的項進行組合得到:
用於中心像素的濾波器係數g中心如下給出:
g中心(σ(梯度),濾波器尺寸(win_size)) = g中心(σ(梯度),4x4)=
分別用於列、行和原始像素的濾波器係數g列、g行和g原始能夠以類似的方式獲得。
聯合濾波器設計2定義為:
其中,g(.)是銳化分量。繼續,用於中心像素的聯合濾波器是:h中心(梯度, 邊緣角, 5x5) = hEI_ 中心(梯度, 邊緣角, 4x4)+ g中心(σ(梯度), 4x4),其中,hEI_ 中心與等式(8)中定義的相同。
在聯合濾波器的這兩個實施方式(設計1和設計2)中使用的濾波器的總數量與等式(13)中給出的相同。這兩個方法中的濾波器尺寸在表7中列出。設計2使用較少的濾波器,從而產生較低的複雜度。設計2還可以比設計1提供更好的主觀品質,並且其具有與兩級濾波器方法類似的品質。這是預料到的,因為設計2在銳化期間使用所估計的被內插的像素,因此在數學上接近近似於兩級濾波器方法。已經觀察到,有時設計1提供比設計2更好的主觀品質。因此,可以使用自適應方法來在這兩個聯合濾波器設計之間進行切換;例如,自適應模型可以基於局部影像特性。
表7 針對win_size=5x5的聯合濾波器的尺寸
The interpolation methods and apparatus described herein can be used in wired or wireless networks. A device including a handheld device, a desktop computer, a laptop computer, or other computer can be used to perform the method. This includes cellular phones, PDAs, tablets and/or displays, as well as cable box boxes, televisions, and the like.
The interpolation or video amplification scheme described herein maintains edge fidelity and computational complexity lower than typical FIR filtering techniques. The system can implement any amplification factors N and M that are applied to the width and height of the video, respectively. For simplicity, it is described here that the image or video frame is magnified by a factor of two (N=2 and M=2) in these two dimensions.
Figure 1 shows a pixel grid containing the original pixel (represented by a square) and the estimated pixel or pixel value (indicated by a circle) estimated at the interpolation point. Pixel A has diagonal original pixels; Pixel B has two adjacent original pixels along the same row; and Pixel C has two adjacent original pixels along the same column. Therefore, the estimated pixels can be classified into three groups: (a) pixels having diagonal adjacent original pixels (labeled as "A" in FIG. 1), and estimated pixels located at these interpolation points can be called a central pixel; (b) a neighboring original row pixel as the nearest neighboring pixel (labeled "B" in Figure 1), the estimated pixel may be referred to as a row pixel; and (c) adjacent The original column pixel is the pixel of the nearest neighboring pixel (labeled "C" in Figure 1), which may be referred to as a column pixel.
In one embodiment, the method includes the following aspects: edge detection; edge angle determination; and pixel estimation. In another embodiment, the method can include determining an edge characteristic associated with the interpolation point, wherein the edge characteristic includes an edge size and an edge angle; selecting an interpolation filter in response to the edge angle; and using the selected An interpolation filter is used to determine the pixel value at the interpolation point.
Any software or hardware (collectively referred to as an operator) used to calculate the horizontal and vertical gradients of a pixel can be used for edge detection. In one embodiment, the modified Sobel operator is used because of its low computational complexity. The Sobel operator includes two square masks for calculating the horizontal and vertical gradients. These gradients are then used to calculate or otherwise obtain the angle of the gradient and the angle of the edge or its estimate. The standard Sobel operator is applied to the square pixel grid, and the modified Sobel operator can be applied to the rectangular pixel grid.
When the determined edge characteristics, estimates the interpolated pixel classification or classification points can be used to select different correction Sobel mask G x and G y. As shown in Figures 2A-2C, G x and G y can be used to multiply the image pixels point by point and sum the products to calculate the horizontal (Δx) and vertical (Δy) gradients (or their estimates), respectively. Masked.
In one embodiment, the magnitude G of the gradient (or edge) is determined by calculating G = |Δx| + |Δy|. If G is greater than the critical value T edge , the edge may exist. In the case where G exceeds the critical value, the angle of the edge gradient can be determined as follows. The value of the edge of the threshold T may depend on the resolution of the video, the pixel class, and other factors. Critical values can be determined empirically, and some threshold values that have been found to perform well are given in Table 1 below. The selection of the thresholds can be different: in some embodiments a fixed threshold can be selected, while in other embodiments, the threshold can be adjusted on a frame-by-frame basis or on a pixel block basis. The adaptive threshold can be automatically calculated based on pixel characteristics within the block. At lower resolutions, neighboring pixels are less likely to be coherent, resulting in a larger gradient. Using smaller thresholds will cause many pixels to be classified as edges. Therefore, in order to reduce inaccurate classification, a larger threshold can be used for smaller resolution video.
In various implementations, the interpolation filter can be selected in response to an edge (or gradient) size, an edge (or gradient) angle, or both. In one embodiment, an interpolation filter may be selected for the interpolated point where the corresponding edge size G is less than the T edge . In other embodiments, an interpolation filter may be selected for an interpolation point where the corresponding edge has Δx or Δy equal to zero. In each case, there is no need to determine or provide an edge angle, and the interpolation filter used to insert new pixels at the interpolated point is as follows:
a. For the center pixel: (a0 + a1 + b0 + b1) / 4
b. For row pixels: (a1+b1)/2
c. For column pixels: (b0+b1)/2
Wherein a0, a1, b0 and b1 correspond to the pixels marked in the 2A-2C diagram. Note that Δy=0 and Δx=0 correspond to gradient angles of 0 and 90 degrees, respectively.
[Mu] gradient angle can be calculated as μ = tan -1 (Δy / Δx ). Since the edge angle is perpendicular to the gradient angle, the edge angle is Θ = μ + 90 o . Note that since the edge and edge gradients can be correlated by a simple 90 degree rotation, the two can be used interchangeably if this relationship is considered. It should also be noted that the range of edge (or gradient) angles can be reduced to a range of 0-180 degrees.
In some embodiments, the interpolation filters used at the respective interpolation points can be determined in response to pixel categories and edge angles. In one embodiment, the interpolation assumes that the edges are linear. For curved edges, other angles can be checked during pixel estimation. Alternatively, a Hough transform based method can be used to detect edge pixels belonging to a curve, and these pixels can be used to insert new pixels along the curved edge.
A pixel estimation process for interpolation points will now be described in which the edge size exceeds a critical value. In one embodiment, for the center pixel, the interpolation filter is selected based on whether the edge angle is approximately 45 degrees. One embodiment uses a center interpolation filter for a range of edge angles between 45 and 135 degrees. Alternative embodiments may utilize other angular ranges.
For central pixel estimation, interpolation can be performed as follows:
If θ is between 35 and 55 degrees, the new pixel = (a1 + b0)/2.
If θ is between 125 and 145 degrees, the new pixel = (a0 + b1)/2.
In other cases, the new pixel = (a0 + a1 + b0 + b1) / 4.
For line pixel estimation,
If θ is between 0 and 90 degrees, the new pixel = (b0 + 2 * (a1 + b1) + a2) / 6.
In other cases, the new pixel = (a0 + 2 * (a1 + b1) + b2) / 6.
For column pixel estimation,
If θ is between 0 and 90 degrees, the new pixel = (c0 + 2 *(b0+b1) + a1)/6.
In other cases, the new pixel = (a0 + 2 * (b0 + b1) + c1) / 6.
Thus, the method can utilize an interpolation filter that applies a larger weight to the pixels located in the direction along the edge angle and applies a smaller weight to the direction perpendicular to the edge angle The weight is not applied to the upper pixel. In one embodiment, the interpolation filter applies a larger weight to the nearest neighboring pixel, applies a medium weight to the pixel located in the direction along the edge angle, and applies the smallest weight to be located perpendicular to the edge angle Pixels in the direction or no weight applied to them. In one embodiment, the nearest neighboring pixels are pixels that are in the same column or in the same row as the interpolated points.
In one embodiment, the gradient measurements Δx and Δy can be used with a lookup table or LUT to determine a suitable interpolation filter. The LUT can store the desired filter impulse response, or can simply provide an interpolation filter identifier that can be used to determine and apply a suitable interpolation filter.
One embodiment described herein is a method for interpolating an image, the method comprising: determining an edge characteristic associated with the interpolated point, the edge characteristic including an edge size and an edge angle; selecting the interpolation in response to the edge angle a filter; and using the selected interpolation filter to determine a pixel value at the interpolation point. The edge characteristics may be based on determining a horizontal gradient and a vertical gradient of pixel values in adjacent regions associated with the interpolation points. The adjacent area associated with the interpolation point may be a horizontal rectangular area as shown in FIG. 2B, a vertical rectangular area as shown in FIG. 2C, or a square area as shown in FIG. 2A. In one embodiment, the neighboring region is determined in response to the interpolated point being a column interpolated point, an interpolated point, or a central interpolated point. In some embodiments, a step filter is used (which may be referred to as a "mask" when a convolution operation is not performed) to determine edge characteristics. One such step filter is a modified Sobel operator.
The performance comparison of edge-based interpolation with pentad filter techniques using the video listed in Table 2 is described below. (In comparison, only the quintile filter was used, but no sharpening filter was used.) Both methods were tested by applying these two methods to the transcoded video. The quality of the amplified video is measured using the peak signal to noise ratio (PSNR), video quality metric (VQM) score, and subjective comparison of the luma component. The transcoder's input video is used as a reference for quality measurements here.
A VQM score close to zero indicates no artifacts/damage, while a score close to one indicates severe damage. VQM only intercepted 15 seconds of video for comparison. The amplification scheme was tested on different video and transcoder bit rates, and the results are listed in Table 3. The items "2X" and "3X" in Table 3 refer to code-converted video that are being encoded at half and one-third of the original bit rate, respectively. For both "vid" video, both amplification methods produce low damage, but edge-based interpolation is slightly better than a 5-tap filter. It should be noted that edge-based interpolation looks significantly better when the viewer has two processed video. This is expected because most objective measurement solutions have limitations. Compared to 2X video, 3X video often produces a slightly higher VQM score because lowering the bit rate increases coding distortion.
For Avery's book video, edge-based interpolation has a small perceptual improvement over the 5-tap method compared to the 5-tap method. The VQM score of Avery's book is higher than the VQM score of the "vid" video because the downsampled video is encoded at a bit rate that is 50 times lower than the original video.
For "vid" video, the PSNR is calculated for video segments that do not have scene changes, and the results are reported in Table 4. For Avery's book, consider all frames with a PSNR of 50 dB, as all frames with a higher PSNR may be blank frames. For vid15 and vid16, a maximum PSNR improvement of 0.7 dB is achieved compared to a 5-tap filter. Figure 3 shows a snapshot of the PSNR and frame for the vid01_2X sequence. Obviously, even on a frame-by-frame basis, there is an improvement in PSNR when using an edge-based interpolation method.
In addition to the VQM score, the two amplification methods are also compared by viewing the amplified video. Experimental results obtained by subjecting the video frame to these two amplification schemes show that the edge-based interpolation method has sharper edges and more detail than the 5-tap filter method.
The complexity comparison is performed by running a magnification algorithm on each test sequence of the 1000 frame on a 2.53 GHz PC, and the average coding time of each frame is measured. The relative speed improvement is calculated by obtaining the ratio of the processing time of the 5-tap filter to the edge-based interpolation filter. The results are given in Table 5. We found that edge-based interpolation is up to 2.5 times faster than a 5-tap filter.
Apparatus and methods for interpolating and sharpening video signals are also described herein. In one embodiment, the method is performed by a two-stage filter method in which the interpolation filter amplifies the input video to a higher spatial resolution and then sharpens the filter to enhance the detail in the video. In another embodiment, the method is performed by a joint filter method in which a filter is applied to the input video signal to amplify the input video signal to a higher resolution while enhancing signal details. . The joint filter performs both interpolation and sharpening and can have associated processing gains. In some embodiments, memory access can also be reduced.
As shown in FIG. 4A, two separate filters, an edge-based interpolation filter and an adaptive sharpening filter, are combined to produce an enlarged and sharpened video frame from the original video frame source. An embodiment is depicted in FIG. 4B in which the YUV frame is first magnified twice in two dimensions by using an edge based interpolator. The edge based interpolator described above can be used. The edge information captured during the interpolation of the luminance component (Y) of the frame can be used to determine the pixel to be sharpened and the strength used to control the adaptive sharpening filter. As outlined in FIG. 9, edge information may be obtained from an edge detection filter selected in part at the position of the interpolated pixel as the center pixel A, the column pixel B, or the row pixel C. Sharpening can be performed on the luma component because the luma component usually contains more detail or edges. The sharpened luma component is then combined with the chrominance component to produce an enlarged sharpened frame. Edge information may also be extracted from one or more chrominance components in the chrominance component, and a weighted combination of edge information from Luma and one or more chrominance components from the chrominance components may be used. Additionally, one or more of the chrominance components may also undergo adaptive sharpening.
The block diagrams in Figures 5 and 6 show one algorithm for the edge-based interpolation and sharpening stages, respectively. Two buffers—the mapping [N][M] and the intensity [N][M]—store the sharpening map and edge intensity, respectively, where N and M are the width and height of the enlarged frame. . These two buffers are initialized to zero before interpolating the frame. The item "1" in the map [i][j] indicates that the pixel at (i, j) is to be sharpened, otherwise it is "0". For each pixel (i, j) to be interpolated, a modified Sobel operator is used to calculate the gradient (grad ij ). If grad ij ≧ sharpens the critical value (sharp_thresh), the mapping [i][j] element for the interpolated pixel and its neighboring pixels is set to one (1), where sharp_thresh is the sharpening threshold. Similarly, the intensity [i][j] for the interpolated pixel and its neighboring pixels is set to grad ij . For example, if the gradient of the interpolated pixel at (i, j) is grad ij ≧sharp_thresh, then:
This is illustrated in Figure 8, where the symbol "x" represents the original pixel and the "o" represents the interpolated pixel. Black dots are interpolated edge pixels that are labeled for sharpening, and other pixels in the bounding box are also labeled for sharpening because of their proximity to the edge pixels. In one embodiment, a sharpening threshold of 100 is appropriate, but the selection of the thresholds can be different. In other embodiments, the sharpening threshold may also be determined from local image characteristics and adjusted to different values throughout the image.
Apply Laplacian-Gaussian (LoG) filter to (i,j) mapping [i] after obtaining the interpolated frame and intensity [i][j] and mapping [i][j] [j]=1 pixels. The neighboring pixels of the edge pixels are sharpened to remove artifacts that may appear to be spots due to non-uniform variations in brightness. In the sharpening phase, the mapping [i][j] and intensity [i][j] information from the interpolation phase are used again. This leads to lower computational complexity.
The definition of a continuous form of LoG filter is as follows:
Where x and y are the distances from the center of LoG (ie (0,0)) and σ 2 is the Gaussian variance. Discrete LoG filters of size KxL are defined as follows:
The parameter σ determines the filter strength. A larger σ (> 1) can be used for weak sharpening, while a smaller σ (≦ 1) can be used for stronger sharpening. The sharpening filter can also have various sizes of win_size. For example, for input video with small resolution (< 640x360), you can choose a LoG filter with win_size = 5x5; for higher resolution video (≧ 640x360), you can choose a LoG filter with win_size = 9x9. The filter size determines the number of neighboring pixels that are considered during sharpening. In other embodiments, different filter sizes can be selected based on video resolution and content. In addition, the filter size can be adjusted throughout the image frame. As an example, the operation of a 5x5 LoG filter can be expressed as follows:
Where C[i][j] and C h [i][j] are the amplified pixels at (i, j) and their corresponding high frequency components, respectively. Figure 10 shows the span of pixels to which filter coefficients are applied, where the black dots represent the interpolated pixels (referred to as central pixels) to be sharpened using a 5x5 LoG filter. The LoG filter coefficients are applied to the pixels in the bounding box.
In one embodiment, σ ij is selected based on the edge strengths described below:
Where ρ 1 , . . . , ρ S is the edge strength critical value, and α 1 , . . . , α S are different σ ij values of S associated with different sharpening levels. In one embodiment, σ ij is defined as follows:
In addition to the edge gradient, local image characteristics can also be used to determine σ ij . For example, excessive sharpening of strong edges (ie, using small σ ij values for strong edges) can result in ringing artifacts. To avoid this, a larger σ ij value can be used when the edge intensity [i][j] is increased.
Finally, the sharpened pixel C out [i][j] is obtained as follows:
λ=1 is used in one embodiment, but in other embodiments, different λ values may be used. In other embodiments, the sharpening filter can also be sensitive to noise by detecting noise before the interpolation phase and using this information to select a LoG filter having a larger σ(σ > 1). This will ensure that the noise detected as edges is not amplified. Similarly, in other embodiments, the sharpening process can be made sensitive to spurious edges, such as block artifacts. Block artifact detection algorithms can be used before interpolation. During the sharpening process, pixels classified as pseudo edges can be smoothed rather than sharpened.
Figure 7 depicts an embodiment of a joint edge-based interpolation and adaptive sharpening filter. In the previous two-stage filter method, the interpolated pixels were calculated in the filter window before the sharpening filter was applied. In the joint filter method, the dependency between interpolation and sharpening is reduced. In this embodiment, the edge based interpolation is combined with a sharpening filter to provide a joint filter. A flow chart of the combined filtering method is shown in Fig. 11. The modified Sobel operator is used to calculate the gradient grad ij for the pixel located at (i, j). The use of two buffers - the buffer sharpening flag (sharp_flag) and the future gradient (future_grad) - is used to mark the sharpening of neighboring pixels at (k, n) that have not been interpolated. If the gradient grad ij is greater than or equal to the predefined sharpening threshold sharp_thresh, the following two operations are performed on the adjacent 3x3 pixels, as shown in FIG. 12:
1. Use a LoG filter to sharpen adjacent original pixels, where equation (5) is used to select the σ of the LoG filter based on grad ij . In Fig. 12, these pixels are indicated by "x".
2. Set the sharp_flag and future_grad of neighboring pixels that have not been interpolated as follows:
If the pixel (i, j) is the center pixel, then set:
.
If the pixel (i, j) is a column pixel, then set:
.
If the pixel (i, j) is a row pixel, then set:
.
In one embodiment, sharp_thresh = 100 and the pixel sharpening candidate is a neighboring 3x3 pixel. In other embodiments, a larger neighborhood is considered and a different sharp_thresh can be used; for example, the value of sharp_thresh can be determined based on local image characteristics. If grad ij is smaller than sharp_thresh, it is checked whether the associated sharp_flag[i][j] is equal to 1. If sharp_flag[i][j] is equal to 1, the associated future_grad[i][j] value is used to select the appropriate joint filter, otherwise grad ij is used to select the joint filter. By using future_grad [i] [j], using a gradient system before (ia, jb) at the adjacent pixels are interpolated, wherein grad ia jb ≧ sharp_thresh and 0 ≦ a, b ≦ 1.
As depicted in Figure 12, the black dots are pixels that are being interpolated and sharpened, and have a gradient that is greater than the sharpening threshold. "x" is the original pixel to be sharpened, and the gray point is the pixel to be interpolated and labeled for sharpening. There are three possible scenarios: (a) center, (b) column, and (c) row pixel interpolation and sharpening.
Below are two implementations of the joint filter. In other embodiments, in addition to the original pixels, the joint filter operates in part on the pixels that have been interpolated.
In the two-stage filter method described above, the pixels within win_size are sharpened as shown by the block in FIG. In this embodiment, all of the original pixels for obtaining the interpolated pixels in the bounding box are identified as indicated by the bold X in Fig. 13. The sharpening filter operates on these raw pixels. Since these pixels are used to acquire the interpolated pixels in the bounding box, the high frequency components calculated from these raw pixels will be close to the high frequency components calculated from all the pixels in the bounding box.
As shown below, the joint filter can be written as the sum of two filters:
Where LOG is a sharpening filter applied only to the original pixels, and h EI is an edge-based interpolator. In this embodiment, λ = 1, but in other embodiments different λ values may be selected. The LoG parameter σ is a function of edge strength, such as the function defined in equation (5). For an implementation where the input video resolution is less than 640 x 360, set win_size = 5x5, otherwise set it to 9x9.
The edge-based interpolator h EI is a function of gradient, edge angle, filter size, and pixel class. As described above with reference to FIG. 9, for edge-based interpolation, the interpolated pixels are classified into a center pixel, a column pixel, and a row pixel based on the position of the interpolated pixel in the pixel grid. Therefore, the enlarged frame, including the original pixels, has four pixel categories. Each of the categories of interpolated pixels (eg, center, column, or row) has three filters, each corresponding to one of the three edge directions. For column and row pixels, the three edge directions are no edge, an edge angle between 0 and 90 degrees, and an edge angle between 90 and 180 degrees. For the center pixel, the three edge directions are no edge, an edge angle between 35 and 55 degrees, and an edge angle between 125 and 145 degrees. During sharpening, consider the original pixels used to retrieve the interpolated pixels within win_size. For example, in order to sharpen the center pixel in Fig. 13, filtering is performed on the original pixel of 4x4. Therefore, a 4x4 LOG filter is considered, resulting in a 4x4 joint filter in equation (7). The size of the joint filter is represented by the filter size (win_size), so for Figure 13, the filter size (win_size) = filter size (5x5) = 4x4. Similarly, as shown in Figures 14, 15, and 16, the filter sizes (5x5) for the column, row, and original pixels are 5x4, 4x5, and 5x5, respectively.
In Fig. 13, interpolation and sharpening of the center pixel indicated by black dots are shown. The interpolated pixels within win_size are taken from the original pixels represented as bold "x". A 4x4 LoG filter is applied to these original pixels. In Fig. 14, interpolation and sharpening of column pixels represented by black dots are shown. A 5x4 LoG filter is applied to the original pixel represented by the bold "x". In Fig. 15, interpolation and sharpening of line pixels represented by black dots are shown. A 4x5 LoG filter is applied to the original pixel represented by the bold "x". In Fig. 16, the sharpening of the original pixel labeled "A" is depicted. A 5x5 LoG filter is applied to the original pixel represented by the bold "x".
The following is an implementation of a joint filter for a central pixel of small screen video resolution (eg, 480x204). This program is similar to programs for columns, rows, and center pixels. Small modifications can be made to achieve a joint filter design for greater video resolution (eg, 1080p, 720p). For small screen video resolution (eg using win_size = 5x5), the resulting joint filter size is the filter size (win_size) = 4x4.
The following is an interpolation filter for generating a center pixel based on the edge direction:
Wherein, the edge threshold determines whether the interpolated pixel is an edge pixel.
An embodiment of the central pixel interpolation filter is given as follows:
In one embodiment, the edge threshold > sharp_thresh. A LoG filter having a standard deviation σ of size filter size (win_size) = KxL is as defined in equation (3).
The LoG filters for the four σ values in equation (5) (where α 1 = ∞, α 2 =1, α 3 = 1.2, and α 4 = 1.4) are as follows:
The joint filter for the center pixel is as follows:
Similarly,
When only interpolation is applied and no sharpening is applied, the center of the filter h is used , ∞ . A joint filter for all other pixel classes can be obtained in a similar manner. In one embodiment, four different σ values and two different win_size values listed in equation (5) are used. Therefore, the number of joint filters for the interpolated pixels is:
The original pixels only need to be sharpened, so edge-based interpolation is not included during its filter design. The number of joint filters used for the original pixels is:
The total number of filters required is:
These filters can be stored in a lookup table and can be selected based on pixel class, gradient, edge angle, and win_size.
In an alternative embodiment, the joint filter is the same as equation (7), but it uses a different sharpening component. An example of an alternative joint filter is described for a central pixel of a small screen video resolution (eg, 480x204). The same design approach can be extended to columns, rows, and raw pixels. For this example, win_size = 5x5. Figure 17 shows the original pixels for sharpening and the estimated interpolated pixels, where all pixels within win_size are used for sharpening, and the interpolated pixels (grey dots) are from the original pixels ( Estimated by the bold "x").
In one embodiment, the edge-based interpolation is used to estimate the interpolated pixels (the gray point in Figure 17) when the gradient is less than the edge threshold. This corresponds to the following estimation filters for center, column and row pixels:
Figure 17 can be represented in the form of a table in Table 6 below, where C ij is a pixel. The gray array represents the pixels in win_size = 5x5, and the bold entries represent the original pixels. Therefore, the non-bold pixels in the gray area are the estimated interpolated pixels.
Table 6
The 5x5 sharpening filter is expressed as follows.
A sharpening filter g that operates only on the original pixel but is still capable of providing the same high-frequency component C h as in Equation (16) can be used according to the following formula:
To illustrate the design of g, the coefficients g 00 and g 01 are derived. Each product term in equation (16) is represented by the original pixel as follows:
From equation (17), the sum of the coefficients corresponding to C 00 is g 00 , and the sum of the coefficients corresponding to C 02 is g 01 . Combine the terms in equation (18) to get:
Center filter coefficient g for the center pixel is given as follows:
Center g ([sigma] (gradient), the filter size (win_size)) = g Center ([sigma] (gradient), 4x4) =
The filter coefficients g columns , g rows, and g primitives for column, row, and original pixels, respectively, can be obtained in a similar manner.
The joint filter design 2 is defined as:
Where g(.) is the sharpening component. Continuing, the joint filter for the center pixel is: h center (gradient, edge angle, 5x5) = h EI_ center (gradient, edge angle, 4x4) + g center (σ (gradient), 4x4), where h EI_ The center is the same as defined in equation (8).
The total number of filters used in these two embodiments of the joint filter (Design 1 and Design 2) is the same as given in equation (13). The filter sizes in these two methods are listed in Table 7. Design 2 uses fewer filters, resulting in lower complexity. Design 2 can also provide better subjective quality than Design 1, and it has similar qualities to the two-stage filter approach. This is expected because Design 2 uses the estimated interpolated pixels during sharpening and is therefore mathematically close to the two-stage filter approach. It has been observed that sometimes Design 1 provides better subjective quality than Design 2. Thus, an adaptive method can be used to switch between the two joint filter designs; for example, the adaptive model can be based on local image characteristics.
Table 7 Size of the joint filter for win_size=5x5
這裏描述的內插和銳化濾波器可以被合併到任意的各種終端,諸如但不限於數位電視、無線通信裝置、無線廣播系統、個人數位助理(PDA)、膝上型或桌上型電腦、迷你筆記型電腦、數位相機、數位記錄裝置、視訊遊戲裝置、視訊遊戲操作臺、蜂巢或衛星無線電電話、數位媒體播放器等。
第18A圖示出了可以在其中實現一個或多個所揭露的實施方式的示例性通信系統100的圖示。通信系統100可以是向多個無線用戶提供諸如語音、資料、視訊、訊息發送、廣播等內容的多重存取系統。通信系統100能夠經由系統資源(包括無線頻寬)的共享來使多個無線用戶存取這些內容。例如,通信系統100可以採用一個或多個頻道存取方法,諸如分碼多重存取(CDMA)、分時多重存取(TDMA)、分頻多重存取(FDMA)、正交FDMA(OFDMA)、單載波FDMA(SC-FDMA)等。
如第18A圖所示,通信系統100可以包括無線傳輸/接收單元(WTRU)102a、102b、102c、102d、無線電存取網路(RAN)104、核心網路106、公共切換電話網(PSTN)108、網際網路110和其他網路112,但是應當意識到,所揭露的實施方式涵蓋任意數量的WTRU、基地台、網路及/或網路元件。WTRU 102a、102b、102c、102d中的每一者可以是被配置為在無線環境中操作及/或通信的任意類型的裝置。以示例的方式,WTRU 102a、102b、102c、102d可以被配置為傳送及/或接收無線信號,並且可以包括用戶設備(UE)、行動站、固定或行動用戶單元、呼叫器、蜂巢式電話、個人數位助理(PDA)、智慧型電話、膝上型電腦、迷你筆記型電腦、個人電腦、無線感測器、消費型電子裝置等。
通信系統100還可以包括基地台114a和基地台114b。基地台114a和114b中的每個基地台可以是被配置用於與WTRU 102a、102b、102c、102d中的至少一個WTRU無線介面連接以促進存取一個或多個通信網路(諸如核心網路106、網際網路110及/或網路112)的任意類型的裝置。以示例的方式,基地台114a、114b可以是基地收發站(BTS)、節點B、e節點B、家用節點B、家用e節點B、站點控制器、存取點(AP)、無線路由器等。雖然基地台114a、114b各自被描述為單一元件,但是應當意識到,基地台114a、114b可以包括任意數量的互連基地台及/或網路元件。
基地台114a可以是RAN 104的一部分,該RAN 104還可以包括其他基地台及/或網路元件(未示出),諸如基地台控制器(BSC)、無線電網路控制器(RNC)、中繼節點等。基地台114a及/或基地台114b可以被配置用於在特定的地理區域(其可以成為胞元(未示出))內傳送及/或接收無線信號。胞元還可以被進一步劃分成胞元扇區。例如,與基地台114a相關聯的胞元可以被劃分成三個扇區。因此,在一個實施方式中,基地台114a可以包括三個收發器,即胞元的每個扇區應用一個收發器。在另一實施方式中,基地台114a可以採用多輸入多輸出(MIMO)技術,因此,可以為胞元的每個扇區使用多個收發器。
基地台114a、114b可以經由空氣介面116與WTRU 102a、102b、102c、102d中的一個或多個WTRU進行通信,其中空氣介面116可以是任何適當的無線通信鏈路(例如,射頻(RF)、微波、紅外線(IR)、紫外線(UV)、可見光等)。可以使用任何適當的無線電存取技術(RAT)來建立空氣介面116。
更為具體地,如上所述,通信系統100可以是多重存取系統並且可以採用一個或多個頻道存取方案,諸如CDMA、TDMA、FDMA、OFDMA、SC-FDMA等。例如,RAN 104中的基地台114a和WTRU 102a、102b、102c可以實施諸如通用行動電信系統(UMTS)陸地無線電存取(UTRA)之類的無線電技術,其可以使用寬頻CDMA(WCDMA)來建立空氣介面116。WCDMA可以包括諸如高速封包存取(HSPA)及/或演進HSPA(HSPA+)等通信協定。HSPA可以包括高速下行鏈路封包存取(HSDPA)及/或高速上行鏈路封包存取(HSUPA)。
在另一實施方式中,基地台114a和WTRU 102a、102b、102c可以實施諸如演進UMTS陸地無線電存取(E-UTRA)等無線電技術,其可以使用長期演進(LTE)及/或高級LTE(LTE-A)來建立空氣介面116。
在其他實施方式中,基地台114a和WTRU 102a、102b、102c可以實施諸如IEEE 802.16(即全球微波互通存取(WiMAX))、CDMA2000、CDMA2000 1X、CDMA2000 EV-DO、臨時標準2000(IS-2000)、臨時標準95(IS-95)、臨時標準856(IS-856)、全球行動通信系統(GSM)、增強型資料速率GSM演進(EDGE)、GSM EDGE(GERAN)等無線電技術。
第18A圖中的基地台114b可以是例如無線路由器、家用節點B、家用e節點B或存取點,並且可以利用用於促進局部區域(諸如商業場所、家庭、車輛、校園等)內的無線連接性的任何適當RAT。在一個實施方式中,基地台114b和WTRU 102c、102d可以實現諸如IEEE 802.11的無線電技術,以建立無線區域網路(WLAN)。在另一實施方式中,基地台114b和WTRU 102c、102d可以實現諸如IEEE 802.15的無線電技術,以建立無線個人區域網路(WPAN)。在再一實施方式中,基地台114b和WTRU 102c、102d可以利用基於蜂巢的RAT(例如,WCDMA、CDMA2000、GSM、LTE、LTE-A等)來建立微微胞元或毫微微胞元。如第18A圖所示,基地台114b可以具有到網際網路110的直接連接。因此,基地台114b可以不需要經由核心網路106來存取網際網路110。
RAN 104可以與核心網路106通信,核心網路106可以是被配置用於向WTRU 102a、102b、102c、102d中的一個或多個WTRU提供語音、資料、應用及/或網際網路協定語音(VoIP)服務的任何類型的網路。例如,核心網路106可以提供呼叫控制、計費服務、基於移動定位的服務、預付費呼叫、網際網路連接、視訊分配等及/或執行高階安全功能(諸如用戶認證)。雖然未在第18A圖中示出,但是應當意識到,RAN 104及/或核心網路106可以與其他RAN直接或間接通信,該其他RAN可以採用與RAN 104所採用的RAT相同的RAT或不同的RAT。例如,除了連接到正在利用E-UTRA無線電技術的RAN 104,核心網路106還可以與採用GSM無線電技術的另一RAN(未示出)進行通信。
核心網路106還可以充當閘道,以便WTRU 102a、102b、102c、102d存取PSTN 108、網際網路110及/或其他網路112。PSTN 108可以包括用於提供普通老式電話服務(POTS)的電路交換電話網。網際網路110可以包括使用公共通信協定(諸如TCP/IP網際網路協定組中的傳輸控制協定(TCP)、用戶資料報協定(UDP)和網際網路協定(IP)的全球互連電腦網路和裝置系統。網路112可以包括由其他服務提供者擁有及/或操作的有線或無線通信網路。例如,網路112可以包括連接到一個或多個RAN(其採用與RAN 104所採用的RAT相同或不同的RAT)的另一核心網路。
通信系統100中的一些或所有WTRU 102a、102b、102c、102d可以包括多模式能力,即WTRU 102a、102b、102c、102d可以包括用於經由不同的無線鏈路與不同的無線網路進行通信的多個收發器。例如,第18A圖中所示的WTRU 102c可以被配置為與可以採用基於蜂巢的無線電技術的基地台114a通信、以及與可以採用IEEE 802無線電技術的基地台114b通信。
第18B圖是示例性的WTRU 102的系統圖。如第18B圖所示,WTRU 102可以包括處理器118、收發器120、傳輸/接收元件122、揚聲器/麥克風124、鍵盤126、顯示器/觸控板128、不可移式記憶體130、可移式記憶體132、電源134、全球定位系統(GPS)碼片組136和其他週邊裝置138。應當意識到,在保持與實施方式相一致的同時,WTRU 102可以包括上述元件的任意子組合。
處理器118可以是通用處理器、專用處理器、常規處理器、數位信號處理器(DSP)、多個微處理器、與DSP核相關聯的一或多個微處理器、控制器、微控制器、專用積體電路(ASIC)、現場可編程閘陣列(FPGA)電路、任何其他類型的積體電路(IC)、狀態機等。處理器118可以執行信號編碼、資料處理、功率控制、輸入/輸出處理及/或使WTRU 102能夠在無線環境中操作的任何其他功能。處理器118可以耦合到收發器120,收發器120可以耦合到傳輸/接收元件122。雖然第18圖將處理器118和收發器描繪為分離的元件,但是應當意識到,處理器118和收發器120可以一起集成到電子封裝或晶片中。
傳輸/接收元件122可以被配置用於經由空氣介面116向基地台(例如,基地台114a)傳送信號、或經由空氣介面116接收來自基地台(例如,基地台114a)的信號。例如,在一個實施方式中,傳輸/接收元件122可以是被配置用於傳送及/或接收RF信號的天線。在另一實施方式中,傳輸/接收元件122可以是被配置用於傳送及/或接收例如IR、UV或可見光信號的發光體/偵測器。在再一實施方式中,傳輸/接收元件122可以被配置用於傳送和接收RF和光信號兩者。應當意識到,傳輸/接收元件122可以被配置用於傳送及/或接收無線信號的任何組合。
另外,雖然在第18圖中傳輸/接收元件122被描繪為單一元件,但是WTRU 102可以包括任意數量的傳輸/接收元件122。更具體地,WTRU 102可以採用MIMO技術。因此,在一個實施方式中,WTRU 102可以包括用於經由空氣介面116來傳送和接收無線信號的兩個或更多個傳輸/接收元件122(例如,多個天線)。
收發器120可以被配置用於調變要由傳輸/接收元件122傳送的信號以及解調由傳輸/接收元件122接收到的信號。如上所述,WTRU 102可以具有多模式能力。因此,收發器120可以包括例如用於使WTRU 102能夠經由多個RAT(諸如UTRA和IEEE 802.11)進行通信的多個收發器。
WTRU 102的處理器118可以耦合到並且可以接收來自揚聲器/麥克風124、鍵盤126及/或顯示器/觸控板128(例如,液晶顯示器(LCD)顯示單元或有機發光二級體(OLED)顯示單元)的用戶輸入資料。處理器118還可以向揚聲器/麥克風124、鍵盤126及/或顯示器/觸控板128輸出用戶資料。另外,處理器118可以存取來自任意類型的適當記憶體(諸如不可移式記憶體130及/或可移式記憶體132)的資訊、並且在任意類型的適當記憶體(諸如不可移式記憶體130及/或可移式記憶體132)中儲存資料。不可移式記憶體130可以包括隨機存取記憶體(RAM)、唯讀記憶體(ROM)、硬碟或任意其他類型的記憶體儲存裝置。可移式記憶體132可以包括用戶身份模組(SIM)卡、記憶條、安全數位(SD)記憶卡等。在其他實施方式中,處理器118可以存取來自沒有實體地位於WTRU 102上的記憶體(諸如伺服器或家庭電腦(未示出))的資訊、並且在沒有實體地位於WTRU 102上的記憶體(諸如伺服器或家庭電腦(未示出))中儲存資料。
處理器118可以接收來自電源134的功率、並且可以被配置用於向WTRU 102中的其他元件分配及/或控制功率。電源134可以是用於為WTRU 102供電的任意適當的裝置。例如,電源134可以包括一個或多個乾電池(例如,鎳-鎘(NiCd)、鎳-鋅(NiZn)、鎳金屬氫化物(NiMH)、鋰離子(Li-ion)等)、太陽能電池、燃料電池等。
處理器118還可以耦合到GPS碼片組136,GPS碼片組136可以被配置用於提供關於WTRU 102的目前位置的位置資訊(例如,經度和維度)。除了或替代來自GPS碼片組136的資訊,WTRU 102可以經由空氣介面116從基地台(例如基地台114a、114b)接收位置資訊、及/或基於從兩個或更多個鄰近基地台接收到的信號的時序來確定其位置。應當意識到,在保持與實施方式一致的同時,WTRU 102可以用任何適當的位置確定方法來獲取位置資訊。
處理器118還可以耦合到其他週邊裝置138,週邊裝置138可以包括用於提供其他的特徵、功能及/或有線或無線連接性的一個或多個軟體及/或硬體模組。例如,週邊裝置138可以包括加速計、電子指南針、衛星收發器、數位相機(用於相片或視訊)、通用串列匯流排(USB)埠、振動裝置、電視收發器、免持耳機、藍芽R模組、調頻(FM)無線電單元、數位音樂播放器、媒體播放器、視訊遊戲機模組、網際網路瀏覽器等。
第18C圖是根據實施方式的RAN 104和核心網路106的系統圖。如上所述,RAN 104可以採用UTRA無線電技術來經由空氣介面116與WTRU 102a、102b、102c通信。RAN 104還可以與核心網路106通信。如第18C圖所示,RAN 104可以包括節點B 140a、104b、140c,其中節點B 140a、104b、140c的每個可以包括用於經由空氣介面116與WTRU 102a、102b、102c通信的一個或多個收發器。節點B 140a、104b、140c每個可以與RAN 104中的特定胞元(未示出)相關聯。RAN 104還可以包括RNC 142a、142b。應當意識到,在保持與實施方式相一致的同時,RAN 104可以包括任意數量的節點B和RNC。
如第18C圖所示,節點B 140a、140b可以與RNC 142a通信。另外,節點B 140c可以與RNC 142b通信。節點B 140a、140b、140c可以經由Iub介面與各自的RNC 142a、142b通信。RNC 142a、142b可以經由Iur介面彼此進行通信。RNC 142a、142b中的每一者被配置用於控制其所連接到的各自的節點B 140a、140b、140c。另外,RNC 142a、142b中的每一者可以被配置用於執行或支援其他功能,諸如外環功率控制、負載控制、准許控制、封包排程、切換控制、巨集分集、安全功能、資料加密等。
第18C圖中所示的核心網路106可以包括媒體閘道(MGW)144、行動交換中心(MSC)146、服務GPRS支援節點(SGSN)148、及/或閘道GPRS支援節點(GGSN)150。雖然前述元件的每個元件被描繪為核心網路106的一部分,但是應當意識到,這些元件中的任何一個元件都可以由除了核心網路操作者之外的實體所擁有及/或操作。
RAN 104中的RNC 142a可以經由IuCS介面連接到核心網路106中的MSC 146。MSC 146可以連接到MGW 144。MSC 146和MGW 144可以向WTRU 102a、102b、102c提供到電路交換網(諸如PSTN 108)的存取,以促進WTRU 102a、102b、102c與傳統陸線通信裝置之間的通信。
RAN 104中的RNC 142a還可以經由IuPS介面連接到核心網路106中的SGSN 148。SGSN 148可以連接到GGSN 150。SGSN 148和GGSN 150可以向WTRU 102a、102b、102c提供到封包交換網路(諸如網際網路110)的存取,以促進WTRU 102a、102b、102c與IP賦能的裝置之間的通信。
如上所述,核心網路106還可以連接到網路112,網路112可以包括由其他服務提供者所擁有及/或操作的其他有線或無線網路。
第18D圖是根據實施方式的RAN 104和核心網路106的系統圖。如上所述,RAN 104可以採用E-UTRA無線電技術來經由空氣介面116與WTRU 102a、102b、102c通信。RAN 104還可以與核心網路106通信。
RAN 104可以包括e節點B 160a、160b、160c,但是應當意識到,在保持與實施方式的一致性的同時,RAN 104可以包括任何數量的e節點B。e節點B 160a、160b、160c的每個可以包括用於經由空氣介面116與WTRU 102a、102b、120c通信的一個或多個收發器。在一個實施方式中,e節點B可以實施MIMO技術。因此,例如,e節點B 160a可以使用多個天線來向WTRU 102a傳送無線信號和從WTRU 102a接收無線信號。
每個e節點B可以與特定的胞元(未示出)相關聯、並且可以被配置用於處理無線電資源管理決策、切換決策、上行鏈路及/或下行鏈路中的用戶排程等。如第18D圖所示,e節點B可以經由X2介面彼此通信。
第18D圖中所示的核心網路106可以包括移動性管理閘道(MME)162、服務閘道164和封包資料網路(PDN)閘道166。雖然前述元件的每個元件被描繪為核心網路106的一部分,但是應當意識到,這些元件中的任何一個元件都可以由除了核心網路操作者之外的實體所擁有及/或操作。
MME 162可以經由S1介面連接到RAN 104中的e節點B 160a、160b、160c中的每一者。例如,MME 162可以負責認證WTRU 102a、102b、102c的用戶、承載啟動/止動、在WTRU 102a、102b、102c的初始連結期間選擇特定的服務閘道等。MME 162還可以提供用於在RAN 104與採用其他無線電技術(諸如GSM或WCDMA)的RAN(未示出)之間進行切換的控制平面功能。
服務閘道164可以經由S1介面連接到RAN 104中的每個e節點B。服務閘道164通常可以向WTRU 102a、102b、102c路由和轉發用戶資料封包、以及路由和轉發來自WTRU 102a、102b、102c的用戶資料封包。服務閘道164還可以執行其他功能,諸如在e節點B間切換期間錨定用戶平面、在下行鏈路資料可由WTRU 102a、102b、102c使用時觸發傳呼、管理和儲存WTRU 102a、102b、102c的上下文等。
服務閘道164還可以連接到PDN閘道166以促進WTRU 102a、102b、102c與IP賦能的裝置之間的通信,其中PDN閘道166可以向WTRU 102a、102b、102c提供到封包交換網路(諸如網際網路110)的存取。
核心網路106可以促進與其他網路的通信。例如,核心網路106可以向WTRU 102a、102b、102c提供到電路交換網(諸如PSTN 108)的存取,以促進WTRU 102a、102b、102c與傳統陸線通信裝置之間的通信。例如,核心網路106可以包括或者可以與IP閘道(例如,IP多媒體子系統(IMS)伺服器)通信,其中IP閘道充當核心網路106與PSTN 108之間的介面。另外,核心網路106可以向WTRU 102a、102b、102c提供到網路112的存取,其中網路112可以包括由其他服務提供者所擁有及/或操作的其他有線或無線網路。
第18E圖是根據實施方式的RAN 104和核心網路106的系統圖。RAN 104可以是採用IEEE 802.16無線電技術來經由空氣介面116與WTRU 102a、102b、102c通信的存取服務網路(ASN)。如下面將進一步討論的,WTRU 102a、102b、102c、RAN 104和核心網路106的不同功能實體之間的通信鏈路可以被定義為參考點。
如第18E圖所示,RAN 104可以包括基地台170a、170b、170c和ASN閘道172,但是應當意識到,在保持與實施方式的一致性的同時,RAN 104可以包括任意數量的基地台和ASN閘道。基地台170a、170b、170c每個可以與RAN 104中的特定胞元(未示出)相關聯、並且各可以包括用於經由空氣介面116與WTRU 102a、102b、102c通信的一個或多個收發器。在一個實施方式中,基地台170a、170b、170c可以實現MIMO技術。因此,例如,基地台170a可以使用多個天線來向WTRU 102a傳送無線信號和從WTRU 102a接收無線信號。基地台170a、170b、170c還可以提供移動性管理功能,諸如切換觸發、隧道建立、無線電資源管理、訊務分類、服務品質(QoS)策略執行等。ASN閘道172可以充當訊務聚合點、並且可以負責傳呼、用戶特性檔快取、到核心網路106的路由等。
WTRU 102a、102b、102c與RAN 104之間的空氣介面116可以被定義為用於實現IEEE 802.16規範的R1參考點。另外,每個WTRU 102a、102b、102c都可以建立與核心網路106的邏輯介面(未示出)。WTRU 102a、102b、102c與核心網路106之間的邏輯介面可以被定義為R2參考點,其可以用於認證、授權、IP主機配置管理及/或移動性管理。
每個基地台140a、140b、140c之間的通信鏈路可以被定義為R8參考點,其包括用於促進WTRU切換和基地台之間的資料傳遞的協定。基地台與ASN閘道172之間的通信鏈路可以被定義為R6參考點。R6參考點可以包括用於促進基於與每個WTRU 102a、102b、102c相關聯的移動性事件的移動性管理的協定。
如第18C圖所示,RAN 104可以連接到核心網路106。RAN 104與核心網路106之間的通信鏈路可以被定義為R3參考點,其包括用於促進例如資料傳遞和移動性管理能力的協定。核心網路106可以包括行動IP本地代理(MIP-HA)144、認證、授權、記賬(AAA)伺服器176和閘道178。雖然前述元件的每個元件被描繪為核心網路106的一部分,但是應當意識到,這些元件中的任何一個元件都可以由除了核心網路操作者之外的實體所擁有及/或操作。
MIP-HA可以負責IP位址管理、並且能夠使WTRU 102a、102b、102c在不同的ASN及/或不同的核心網路之間漫遊。MIP-HA 174可以向WTRU 102a、102b、102c提供到封包交換網路(諸如網際網路110)的存取,以促進WTRU 102a、102b、102c與IP賦能的裝置之間的通信。AAA伺服器146可以負責用戶認證和用於支援用戶服務。閘道178可以促進與其他網路的互連。例如,閘道178可以向WTRU 102a、102b、102c提供到電路交換網(諸如PSTN 108)的存取,以促進WTRU 102a、102b、102c與傳統陸線通信裝置的通信。另外,閘道178可以向WTRU 102a、102b、102c提供到網路112的存取,其中網路112可以包括由其他服務提供者所擁有及/或操作的其他有線或無線網路。
雖然未在第18E圖中示出,但是應當意識到,RAN 104可以連接到其他ASN,並且核心網路106可以連接到其他核心網路。RAN 104與其他ASN之間的通信鏈路可以被定義為R4參考點,其可以包括用於協調RAN 104與其他ASN之間的WTRU 102a、102b、102c的移動性的協定。核心網路106與其他核心網路之間的通信鏈路可以被定義為R5參考點,其可以包括用於促進本地核心網路與被訪問的核心網路之間的互連的協定。
雖然上面在特定的組合中描述了特徵和元素,但是本領域中具有通常知識者應當意識到,每個特徵或元素都可以被單獨使用或者可以用與其他特徵和元素的任意組合的方式使用。另外,這裏描述的方法可以在被包括在電腦或處理器所執行的電腦可讀媒體中的電腦程式、軟體或韌體中實現。電腦可讀媒體的示例包括電腦可讀儲存媒體。電腦可讀儲存媒體的示例包括但不限於唯讀記憶體(ROM)、隨機儲存記憶體(RAM)、暫存器、快取記憶體、半導體儲存裝置、諸如內部硬碟或可移動盤之類的磁性媒體、磁光媒體和例如CD-ROM盤和數位多功能盤(DVD)之類的光學媒體。與軟體相關聯的處理器可以用於實現在WTRU、UE、終端、基地台、RNC或任何主電腦中使用的射頻收發器。
在不背離本發明範圍的情況下,上面描述的各種方法、裝置和系統的變形是可能的。鑒於能夠被應用的各種實施方式,應當理解,所示出的實施方式僅是示例性,並且不應當被認為是限制所附申請專利範圍書的範圍。
而且,在上面所述的實施方式中,指出了包含處理器的處理平臺、計算系統、控制器和其他裝置。這些裝置可以包含至少一個中央處理單元(“CPU”)和記憶體。根據電腦程式領域內的技術人員的實務,操作或指令的動作和符號表示可以由各種CPU和記憶體執行。這種動作和操作或指令可以被稱為正在“被執行”、“被電腦執行”或“被CPU執行”。
本領域中具有通常知識者將意識到,動作和符號表示的操作或指令可以包括由CPU執行的電子信號的操縱。電子系統表示一種資料位元,該資料位元能夠促使電子信號的有效變換或降低、和記憶體系統中的記憶體位置處的資料位元的維護,從而重新配置或以其他方式改變CPU的操作以及信號的其他處理。資料位元被保持於其中的記憶體位置是具有與資料位元相對應的或表示資料位元的特定電、磁、光或有機特性的實體位置。應當理解,示例性的實施方式並不侷限於上面提到的平臺或CPU,並且其他平臺和CPU也可以支援上面描述的方法。
資料位元還可以被保持在電腦可讀媒體上,其中電腦可讀媒體包括可由CPU讀取的磁片、光碟和任意其他揮發性(例如,隨機存取記憶體(“RAM”)或非揮發性(例如,唯讀記憶體(“ROM”)大容量儲存系統。電腦可讀媒體可以包括協作或互連的電腦可讀媒體,其排他性地存在於處理系統上或者在多個互連的處理系統(其可以位於處理系統本地或遠離處理系統)之間進行分佈。應當理解,示例性的實施方式並不限於上面提到的記憶體,並且其他平臺和記憶體也可以支援上面描述的方法。
描述本申請時所使用的任何元素、動作或指令都不應當被解釋為對本發明是關鍵或必不可少的,除非以此方式明確地進行了描述。而且,如本文所使用的,單詞“一”意欲包括一個或多個項目。在期望僅指代一個項目時,使用術語“一個(one)”或類似語言。另外,如本文所使用的,多個項目及/或多個類別的項目中的任一個意欲包括單獨的這些項目及/或這些類別的項目或與其他項目及/或其他類別的項目相結合中的任一個、任意組合、任意多個及/或任意多個組合。另外,如本文所使用的,術語“集合”意欲包括任意數量的項目,包括零。另外,如本文所使用的,術語“數量”意欲包括任意數量,包括零。
而且,申請專利範圍不應當被解釋為局限於所描述的次序或元素,除非明確指出。另外,任何申請專利範圍中的術語“裝置(means)”意欲援引35 U.S.C. §112, ¶ 6,而且並不期望不具有單詞“裝置(means)”的任何申請專利範圍也是如此援引。
The interpolation and sharpening filters described herein can be incorporated into any of a variety of terminals such as, but not limited to, digital televisions, wireless communication devices, wireless broadcast systems, personal digital assistants (PDAs), laptops or desktop computers, Mini notebook computers, digital cameras, digital recording devices, video game devices, video game consoles, cellular or satellite radio phones, digital media players, etc.
FIG. 18A illustrates an illustration of an exemplary communication system 100 in which one or more disclosed embodiments may be implemented. Communication system 100 may be a multiple access system that provides content, such as voice, material, video, messaging, broadcast, etc., to multiple wireless users. Communication system 100 is capable of enabling multiple wireless users to access such content via sharing of system resources, including wireless bandwidth. For example, communication system 100 can employ one or more channel access methods, such as code division multiple access (CDMA), time division multiple access (TDMA), frequency division multiple access (FDMA), orthogonal FDMA (OFDMA). Single carrier FDMA (SC-FDMA), etc.
As shown in FIG. 18A, communication system 100 can include wireless transmit/receive units (WTRUs) 102a, 102b, 102c, 102d, radio access network (RAN) 104, core network 106, public switched telephone network (PSTN). 108. Internet 110 and other networks 112, but it should be appreciated that the disclosed embodiments encompass any number of WTRUs, base stations, networks, and/or network elements. Each of the WTRUs 102a, 102b, 102c, 102d may be any type of device configured to operate and/or communicate in a wireless environment. By way of example, the WTRUs 102a, 102b, 102c, 102d may be configured to transmit and/or receive wireless signals, and may include user equipment (UE), mobile stations, fixed or mobile subscriber units, pagers, cellular phones, Personal digital assistants (PDAs), smart phones, laptops, mini-notebooks, personal computers, wireless sensors, consumer electronics, and more.
Communication system 100 can also include a base station 114a and a base station 114b. Each of the base stations 114a and 114b may be configured to wirelessly interface with at least one of the WTRUs 102a, 102b, 102c, 102d to facilitate access to one or more communication networks (such as a core network) 106. Any type of device of the Internet 110 and/or the network 112). By way of example, base stations 114a, 114b may be base transceiver stations (BTS), node B, eNodeB, home node B, home eNodeB, site controller, access point (AP), wireless router, etc. . While base stations 114a, 114b are each depicted as a single component, it should be appreciated that base stations 114a, 114b can include any number of interconnected base stations and/or network elements.
The base station 114a may be part of the RAN 104, which may also include other base stations and/or network elements (not shown), such as a base station controller (BSC), a radio network controller (RNC), Following the node and so on. Base station 114a and/or base station 114b may be configured to transmit and/or receive wireless signals within a particular geographic area (which may be a cell (not shown)). The cell can also be further divided into cell sectors. For example, a cell associated with base station 114a can be divided into three sectors. Thus, in one embodiment, base station 114a may include three transceivers, i.e., one transceiver per sector of the cell. In another embodiment, the base station 114a may employ multiple input multiple output (MIMO) technology, and thus, multiple transceivers may be used for each sector of the cell.
The base stations 114a, 114b can communicate with one or more of the WTRUs 102a, 102b, 102c, 102d via the air interface 116, where the air interface 116 can be any suitable wireless communication link (eg, radio frequency (RF), Microwave, infrared (IR), ultraviolet (UV), visible light, etc.). The air interface 116 can be established using any suitable radio access technology (RAT).
More specifically, as noted above, communication system 100 can be a multiple access system and can employ one or more channel access schemes such as CDMA, TDMA, FDMA, OFDMA, SC-FDMA, and the like. For example, base station 114a and WTRUs 102a, 102b, 102c in RAN 104 may implement a radio technology such as Universal Mobile Telecommunications System (UMTS) Terrestrial Radio Access (UTRA), which may use broadband CDMA (WCDMA) to establish air. Interface 116. WCDMA may include communication protocols such as High Speed Packet Access (HSPA) and/or Evolution HSPA (HSPA+). HSPA may include High Speed Downlink Packet Access (HSDPA) and/or High Speed Uplink Packet Access (HSUPA).
In another embodiment, base station 114a and WTRUs 102a, 102b, 102c may implement radio technologies such as Evolved UMTS Terrestrial Radio Access (E-UTRA), which may use Long Term Evolution (LTE) and/or LTE-Advanced (LTE). -A) to establish the air interface 116.
In other embodiments, base station 114a and WTRUs 102a, 102b, 102c may implement, for example, IEEE 802.16 (ie, Worldwide Interoperability for Microwave Access (WiMAX)), CDMA2000, CDMA2000 1X, CDMA2000 EV-DO, Provisional Standard 2000 (IS-2000) Radio Technologies such as Interim Standard 95 (IS-95), Interim Standard 856 (IS-856), Global System for Mobile Communications (GSM), Enhanced Data Rate GSM Evolution (EDGE), GSM EDGE (GERAN).
The base station 114b in FIG. 18A may be, for example, a wireless router, a home Node B, a home eNodeB, or an access point, and may be utilized to facilitate wireless within a local area (such as a business location, home, vehicle, campus, etc.) Any suitable RAT for connectivity. In one embodiment, base station 114b and WTRUs 102c, 102d may implement a radio technology such as IEEE 802.11 to establish a wireless local area network (WLAN). In another embodiment, base station 114b and WTRUs 102c, 102d may implement a radio technology such as IEEE 802.15 to establish a wireless personal area network (WPAN). In still another embodiment, base station 114b and WTRUs 102c, 102d may utilize a cellular based RAT (eg, WCDMA, CDMA2000, GSM, LTE, LTE-A, etc.) to establish picocells or femtocells. As shown in FIG. 18A, base station 114b may have a direct connection to internet 110. Thus, base station 114b may not need to access Internet 110 via core network 106.
The RAN 104 can communicate with a core network 106, which can be configured to provide voice, data, application, and/or internet protocol voice to one or more of the WTRUs 102a, 102b, 102c, 102d. Any type of network (VoIP) service. For example, core network 106 may provide call control, billing services, mobile location based services, prepaid calling, internet connectivity, video distribution, etc. and/or perform high level security functions such as user authentication. Although not shown in FIG. 18A, it should be appreciated that the RAN 104 and/or the core network 106 can communicate directly or indirectly with other RANs that can employ the same RAT or different RAT as the RAN 104 employs. RAT. For example, in addition to being connected to the RAN 104 that is utilizing the E-UTRA radio technology, the core network 106 can also communicate with another RAN (not shown) employing a GSM radio technology.
The core network 106 can also act as a gateway for the WTRUs 102a, 102b, 102c, 102d to access the PSTN 108, the Internet 110, and/or other networks 112. The PSTN 108 may include a circuit switched telephone network for providing Plain Old Telephone Service (POTS). The Internet 110 may include a globally interconnected computer network that uses public communication protocols such as Transmission Control Protocol (TCP), User Datagram Protocol (UDP), and Internet Protocol (IP) in the TCP/IP Internet Protocol Group. Road and device system. Network 112 may include a wired or wireless communication network that is owned and/or operated by other service providers. For example, network 112 may include a connection to one or more RANs (which are employed with RAN 104) Another core network of the same or different RATs of the RAT.
Some or all of the WTRUs 102a, 102b, 102c, 102d in the communication system 100 may include multi-mode capabilities, i.e., the WTRUs 102a, 102b, 102c, 102d may include communications for communicating with different wireless networks via different wireless links. Multiple transceivers. For example, the WTRU 102c shown in FIG. 18A can be configured to communicate with a base station 114a that can employ a cellular-based radio technology, and with a base station 114b that can employ an IEEE 802 radio technology.
Figure 18B is a system diagram of an exemplary WTRU 102. As shown in FIG. 18B, the WTRU 102 may include a processor 118, a transceiver 120, a transmit/receive element 122, a speaker/microphone 124, a keyboard 126, a display/touchpad 128, a non-removable memory 130, and a removable Memory 132, power source 134, global positioning system (GPS) chip set 136, and other peripheral devices 138. It should be appreciated that the WTRU 102 may include any sub-combination of the above-described elements while remaining consistent with the embodiments.
The processor 118 can be a general purpose processor, a special purpose processor, a conventional processor, a digital signal processor (DSP), a plurality of microprocessors, one or more microprocessors associated with the DSP core, a controller, a micro control , dedicated integrated circuit (ASIC), field programmable gate array (FPGA) circuit, any other type of integrated circuit (IC), state machine, etc. The processor 118 can perform signal coding, data processing, power control, input/output processing, and/or any other functionality that enables the WTRU 102 to operate in a wireless environment. The processor 118 can be coupled to a transceiver 120 that can be coupled to the transmit/receive element 122. Although FIG. 18 depicts processor 118 and transceiver as separate components, it should be appreciated that processor 118 and transceiver 120 can be integrated together into an electronic package or wafer.
The transmit/receive element 122 can be configured to transmit signals to a base station (e.g., base station 114a) via air interface 116 or to receive signals from a base station (e.g., base station 114a) via air interface 116. For example, in one embodiment, the transmit/receive element 122 can be an antenna configured to transmit and/or receive RF signals. In another embodiment, the transmit/receive element 122 can be an illuminant/detector configured to transmit and/or receive, for example, IR, UV, or visible light signals. In still another embodiment, the transmit/receive element 122 can be configured to transmit and receive both RF and optical signals. It should be appreciated that the transmit/receive element 122 can be configured to transmit and/or receive any combination of wireless signals.
Additionally, although transmission/reception element 122 is depicted as a single element in FIG. 18, WTRU 102 may include any number of transmission/reception elements 122. More specifically, the WTRU 102 may employ MIMO technology. Thus, in one embodiment, the WTRU 102 may include two or more transmission/reception elements 122 (e.g., multiple antennas) for transmitting and receiving wireless signals via the air interface 116.
The transceiver 120 can be configured to modulate signals to be transmitted by the transmission/reception element 122 and to demodulate signals received by the transmission/reception element 122. As noted above, the WTRU 102 may have multi-mode capabilities. Thus, transceiver 120 may include, for example, multiple transceivers for enabling WTRU 102 to communicate via multiple RATs, such as UTRA and IEEE 802.11.
The processor 118 of the WTRU 102 can be coupled to and can receive from the speaker/microphone 124, the keyboard 126, and/or the display/touchpad 128 (eg, a liquid crystal display (LCD) display unit or an organic light emitting diode (OLED) display unit) ) User input data. The processor 118 can also output user profiles to the speaker/microphone 124, the keyboard 126, and/or the display/touchpad 128. In addition, the processor 118 can access information from any type of suitable memory (such as the non-removable memory 130 and/or the removable memory 132) and in any type of appropriate memory (such as non-removable memory). The data is stored in the body 130 and/or the removable memory 132). The non-removable memory 130 may include random access memory (RAM), read only memory (ROM), hard disk, or any other type of memory storage device. The removable memory 132 can include a Subscriber Identity Module (SIM) card, a memory stick, a secure digital (SD) memory card, and the like. In other embodiments, the processor 118 may access information from memory (such as a server or home computer (not shown)) that is not physically located on the WTRU 102, and is not physically located on the WTRU 102. The data is stored in a body such as a server or a home computer (not shown).
The processor 118 can receive power from the power source 134 and can be configured to allocate and/or control power to other elements in the WTRU 102. Power source 134 may be any suitable device for powering WTRU 102. For example, the power source 134 may include one or more dry cells (eg, nickel-cadmium (NiCd), nickel-zinc (NiZn), nickel metal hydride (NiMH), lithium ion (Li-ion), etc.), solar cells, fuel Battery, etc.
The processor 118 may also be coupled to a set of GPS chips 136 that may be configured to provide location information (e.g., longitude and dimension) with respect to the current location of the WTRU 102. In addition to or in lieu of information from the GPS chipset 136, the WTRU 102 may receive location information from a base station (e.g., base station 114a, 114b) via air interface 116, and/or based on receiving from two or more neighboring base stations. The timing of the signal to determine its position. It should be appreciated that the WTRU 102 may obtain location information using any suitable location determination method while remaining consistent with the implementation.
The processor 118 can also be coupled to other peripheral devices 138, which can include one or more software and/or hardware modules for providing other features, functionality, and/or wired or wireless connectivity. For example, peripheral device 138 may include an accelerometer, an electronic compass, a satellite transceiver, a digital camera (for photo or video), a universal serial bus (USB) port, a vibrating device, a television transceiver, a hands-free headset, and a Bluetooth device. R module, FM radio unit, digital music player, media player, video game console module, internet browser, etc.
Figure 18C is a system diagram of RAN 104 and core network 106, in accordance with an embodiment. As noted above, the RAN 104 can employ UTRA radio technology to communicate with the WTRUs 102a, 102b, 102c via the air interface 116. The RAN 104 can also communicate with the core network 106. As shown in FIG. 18C, RAN 104 may include Node Bs 140a, 104b, 140c, wherein each of Node Bs 140a, 104b, 140c may include one or more for communicating with WTRUs 102a, 102b, 102c via air interface 116. Transceivers. Node Bs 140a, 104b, 140c may each be associated with a particular cell (not shown) in RAN 104. The RAN 104 may also include RNCs 142a, 142b. It should be appreciated that the RAN 104 may include any number of Node Bs and RNCs while remaining consistent with the implementation.
As shown in Figure 18C, Node Bs 140a, 140b can communicate with RNC 142a. Additionally, Node B 140c can communicate with RNC 142b. Node Bs 140a, 140b, 140c can communicate with respective RNCs 142a, 142b via an Iub interface. The RNCs 142a, 142b can communicate with each other via the Iur interface. Each of the RNCs 142a, 142b is configured to control the respective Node Bs 140a, 140b, 140c to which it is connected. Additionally, each of the RNCs 142a, 142b can be configured to perform or support other functions, such as outer loop power control, load control, admission control, packet scheduling, handover control, macro diversity, security functions, data encryption. Wait.
The core network 106 shown in FIG. 18C may include a media gateway (MGW) 144, a mobile switching center (MSC) 146, a serving GPRS support node (SGSN) 148, and/or a gateway GPRS support node (GGSN) 150. . While each element of the aforementioned elements is depicted as being part of core network 106, it should be appreciated that any of these elements can be owned and/or operated by entities other than the core network operator.
The RNC 142a in the RAN 104 can be connected to the MSC 146 in the core network 106 via an IuCS interface. The MSC 146 can be connected to the MGW 144. The MSC 146 and the MGW 144 may provide the WTRUs 102a, 102b, 102c with access to a circuit-switched network, such as the PSTN 108, to facilitate communication between the WTRUs 102a, 102b, 102c and conventional landline communication devices.
The RNC 142a in the RAN 104 can also be connected to the SGSN 148 in the core network 106 via an IuPS interface. The SGSN 148 can be connected to the GGSN 150. The SGSN 148 and GGSN 150 may provide the WTRUs 102a, 102b, 102c with access to a packet switched network, such as the Internet 110, to facilitate communication between the WTRUs 102a, 102b, 102c and IP-enabled devices.
As noted above, the core network 106 can also be connected to the network 112, which can include other wired or wireless networks that are owned and/or operated by other service providers.
Figure 18D is a system diagram of RAN 104 and core network 106, in accordance with an embodiment. As noted above, the RAN 104 can employ E-UTRA radio technology to communicate with the WTRUs 102a, 102b, 102c via the air interface 116. The RAN 104 can also communicate with the core network 106.
The RAN 104 may include eNodeBs 160a, 160b, 160c, but it should be appreciated that the RAN 104 may include any number of eNodeBs while maintaining consistency with the embodiments. Each of the eNodeBs 160a, 160b, 160c may include one or more transceivers for communicating with the WTRUs 102a, 102b, 120c via the air interface 116. In one embodiment, the eNodeB can implement MIMO technology. Thus, for example, the eNodeB 160a may use multiple antennas to transmit wireless signals to and receive wireless signals from the WTRU 102a.
Each eNodeB can be associated with a particular cell (not shown) and can be configured to handle radio resource management decisions, handover decisions, user scheduling in the uplink and/or downlink, and the like. As shown in Fig. 18D, the eNodeBs can communicate with each other via the X2 interface.
The core network 106 shown in FIG. 18D may include a mobility management gateway (MME) 162, a service gateway 164, and a packet data network (PDN) gateway 166. While each element of the aforementioned elements is depicted as being part of core network 106, it should be appreciated that any of these elements can be owned and/or operated by entities other than the core network operator.
The MME 162 may be connected to each of the eNodeBs 160a, 160b, 160c in the RAN 104 via an S1 interface. For example, MME 162 may be responsible for authenticating users of WTRUs 102a, 102b, 102c, bearer activation/deactivation, selecting a particular service gateway during initial connection of WTRUs 102a, 102b, 102c, and the like. The MME 162 may also provide control plane functionality for switching between the RAN 104 and a RAN (not shown) employing other radio technologies, such as GSM or WCDMA.
Service gateway 164 may be connected to each eNodeB in RAN 104 via an S1 interface. The service gateway 164 can generally route and forward user data packets to the WTRUs 102a, 102b, 102c, and route and forward user data packets from the WTRUs 102a, 102b, 102c. The service gateway 164 may also perform other functions, such as anchoring the user plane during handover between eNodeBs, triggering paging, managing and storing the WTRUs 102a, 102b, 102c when the downlink profiles are available to the WTRUs 102a, 102b, 102c. Context, etc.
The service gateway 164 can also be coupled to the PDN gateway 166 to facilitate communication between the WTRUs 102a, 102b, 102c and IP-enabled devices, wherein the PDN gateway 166 can provide the WTRUs 102a, 102b, 102c to the packet switched network. Access (such as the Internet 110).
The core network 106 can facilitate communication with other networks. For example, core network 106 may provide WTRUs 102a, 102b, 102c with access to a circuit-switched network, such as PSTN 108, to facilitate communication between WTRUs 102a, 102b, 102c and conventional landline communication devices. For example, core network 106 may include or may be in communication with an IP gateway (eg, an IP Multimedia Subsystem (IMS) server), where the IP gateway acts as an interface between core network 106 and PSTN 108. In addition, core network 106 can provide WTRUs 102a, 102b, 102c with access to network 112, which can include other wired or wireless networks that are owned and/or operated by other service providers.
Figure 18E is a system diagram of RAN 104 and core network 106, in accordance with an embodiment. The RAN 104 may be an Access Service Network (ASN) that communicates with the WTRUs 102a, 102b, 102c via the air interface 116 using IEEE 802.16 radio technology. As discussed further below, the communication links between the different functional entities of the WTRUs 102a, 102b, 102c, RAN 104, and core network 106 may be defined as reference points.
As shown in FIG. 18E, the RAN 104 may include base stations 170a, 170b, 170c and ASN gateway 172, but it should be appreciated that the RAN 104 may include any number of base stations and while maintaining consistency with the embodiments. ASN gateway. Base stations 170a, 170b, 170c may each be associated with a particular cell (not shown) in RAN 104 and may each include one or more transceivers for communicating with WTRUs 102a, 102b, 102c via air interface 116. Device. In one embodiment, base stations 170a, 170b, 170c may implement MIMO technology. Thus, for example, base station 170a can use multiple antennas to transmit wireless signals to, and receive wireless signals from, WTRU 102a. The base stations 170a, 170b, 170c may also provide mobility management functions such as handover triggering, tunnel establishment, radio resource management, traffic classification, quality of service (QoS) policy enforcement, and the like. The ASN gateway 172 can act as a traffic aggregation point and can be responsible for paging, user profile cache, routing to the core network 106, and the like.
The air interface 116 between the WTRUs 102a, 102b, 102c and the RAN 104 may be defined as an Rl reference point for implementing the IEEE 802.16 specification. In addition, each of the WTRUs 102a, 102b, 102c can establish a logical interface (not shown) with the core network 106. The logical interface between the WTRUs 102a, 102b, 102c and the core network 106 can be defined as an R2 reference point that can be used for authentication, authorization, IP host configuration management, and/or mobility management.
The communication link between each of the base stations 140a, 140b, 140c may be defined as an R8 reference point that includes protocols for facilitating data transfer between the WTRU handover and the base station. The communication link between the base station and the ASN gateway 172 can be defined as an R6 reference point. The R6 reference point may include an agreement to facilitate mobility management based on mobility events associated with each of the WTRUs 102a, 102b, 102c.
As shown in FIG. 18C, the RAN 104 can be connected to the core network 106. The communication link between the RAN 104 and the core network 106 can be defined as an R3 reference point that includes protocols for facilitating, for example, data transfer and mobility management capabilities. The core network 106 may include a Mobile IP Home Agent (MIP-HA) 144, an Authentication, Authorization, Accounting (AAA) server 176, and a gateway 178. While each element of the aforementioned elements is depicted as being part of core network 106, it should be appreciated that any of these elements can be owned and/or operated by entities other than the core network operator.
The MIP-HA may be responsible for IP address management and can enable the WTRUs 102a, 102b, 102c to roam between different ASNs and/or different core networks. The MIP-HA 174 may provide the WTRUs 102a, 102b, 102c with access to a packet switched network, such as the Internet 110, to facilitate communication between the WTRUs 102a, 102b, 102c and IP-enabled devices. The AAA server 146 can be responsible for user authentication and for supporting user services. Gateway 178 can facilitate interconnection with other networks. For example, gateway 178 can provide WTRUs 102a, 102b, 102c with access to a circuit-switched network, such as PSTN 108, to facilitate communication of WTRUs 102a, 102b, 102c with conventional landline communication devices. In addition, gateway 178 can provide access to network 112 to WTRUs 102a, 102b, 102c, which can include other wired or wireless networks that are owned and/or operated by other service providers.
Although not shown in Figure 18E, it should be appreciated that the RAN 104 can be connected to other ASNs and the core network 106 can be connected to other core networks. The communication link between the RAN 104 and other ASNs may be defined as an R4 reference point, which may include a protocol for coordinating the mobility of the WTRUs 102a, 102b, 102c between the RAN 104 and other ASNs. The communication link between core network 106 and other core networks may be defined as an R5 reference point, which may include protocols for facilitating interconnection between the local core network and the core network being accessed.
Although features and elements are described above in a particular combination, those of ordinary skill in the art will recognize that each feature or element can be used alone or in any combination with other features and elements. Additionally, the methods described herein can be implemented in a computer program, software or firmware embodied in a computer readable medium embodied by a computer or processor. Examples of computer readable media include computer readable storage media. Examples of computer readable storage media include, but are not limited to, read only memory (ROM), random access memory (RAM), scratchpad, cache memory, semiconductor storage devices, such as internal hard drives or removable disks. Magnetic media, magneto-optical media, and optical media such as CD-ROM discs and digital versatile discs (DVDs). A processor associated with the software can be used to implement a radio frequency transceiver for use in a WTRU, UE, terminal, base station, RNC, or any host computer.
Variations of the various methods, apparatus, and systems described above are possible without departing from the scope of the invention. In view of the various embodiments that can be applied, it is to be understood that the illustrated embodiments are only illustrative and are not to be construed as limiting the scope of the appended claims.
Moreover, in the embodiments described above, a processing platform, computing system, controller, and other devices including a processor are indicated. These devices may include at least one central processing unit ("CPU") and memory. The actions and symbolic representations of operations or instructions can be performed by various CPUs and memories, depending on the practice of the skilled person in the field of computer programming. Such actions and operations or instructions may be referred to as being "executed,""executed by a computer," or "executed by a CPU."
Those of ordinary skill in the art will appreciate that the actions or instructions of the acts and symbolic representations can include manipulation of electronic signals that are executed by the CPU. An electronic system represents a data bit that can cause an effective transformation or reduction of an electronic signal, and maintenance of a data bit at a memory location in a memory system, thereby reconfiguring or otherwise altering the operation of the CPU And other processing of the signal. The memory location in which the data bit is held is the physical location having a particular electrical, magnetic, optical or organic characteristic corresponding to the data bit or representing the data bit. It should be understood that the exemplary embodiments are not limited to the platforms or CPUs mentioned above, and that other platforms and CPUs may also support the methods described above.
The data bits can also be maintained on a computer readable medium, including a magnetic disk, a compact disc, and any other volatiles (eg, random access memory ("RAM") or non-volatile that can be read by the CPU. (eg, a read-only memory ("ROM") mass storage system. Computer-readable media can include cooperating or interconnected computer readable media, either exclusively on a processing system or in multiple interconnected processes The system (which may be local to the processing system or remote from the processing system) is distributed. It should be understood that the exemplary embodiments are not limited to the memory mentioned above, and other platforms and memories may also support the methods described above.
No element, act, or instruction used in the description of the present application should be construed as being critical or essential to the present invention unless specifically described in this manner. Also, as used herein, the word "a" is intended to include one or more items. The term "one" or a similar language is used when it is desired to refer only to one item. In addition, as used herein, any of a plurality of items and/or a plurality of categories of items is intended to include separate items and/or items of these categories or in combination with other items and/or other categories of items. Any one, any combination, any number, and/or any combination. Also, as used herein, the term "set" is intended to include any number of items, including zero. Also, as used herein, the term "amount" is intended to include any number, including zero.
Moreover, the scope of the claims should not be construed as being limited In addition, the term "means" in the scope of any patent application is intended to invoke 35 USC § 112, ¶ 6, and it is not intended that any patent application scope without the word "means" is also invoked.
100...示例性通信系統100. . . Exemplary communication system
102a、102b、102c、102d...無線傳輸/接收單元(WTRU)102a, 102b, 102c, 102d. . . Wireless transmit/receive unit (WTRU)
104...無線電存取網路(RAN)104. . . Radio access network (RAN)
106...核心網路106. . . Core network
108...公共切換電話網(PSTN)108. . . Public Switched Telephone Network (PSTN)
110...網際網路110. . . Internet
112...其他網路112. . . Other network
114a、114b、170a、170b、170c...基地台114a, 114b, 170a, 170b, 170c. . . Base station
116...空氣介面116. . . Air interface
118...處理器118. . . processor
120...收發器120. . . transceiver
122...傳輸/接收元件122. . . Transmission/reception component
124...揚聲器/麥克風124. . . Speaker/microphone
126...鍵盤126. . . keyboard
128...顯示器/觸控板128. . . Display/trackpad
130...不可移式記憶體130. . . Non-removable memory
132...可移式記憶體132. . . Removable memory
134...電源134. . . power supply
136...全球定位系統(GPS)碼片組136. . . Global Positioning System (GPS) chipset
138...週邊裝置138. . . Peripheral device
140a、104b、140c...節點B140a, 104b, 140c. . . Node B
142a、142b...無線電網路控制器(RNC)142a, 142b. . . Radio Network Controller (RNC)
144...媒體閘道(MGW)144. . . Media Gateway (MGW)
146...行動交換中心(MSC)146. . . Mobile Switching Center (MSC)
148...服務GPRS支援節點(SGSN)148. . . Serving GPRS Support Node (SGSN)
150...閘道GPRS支援節點(GGSN)150. . . Gateway GPRS Support Node (GGSN)
160a、160b、160c...e節點B160a, 160b, 160c. . . eNodeB
162...移動性管理閘道(MME)162. . . Mobility Management Gateway (MME)
164...服務閘道164. . . Service gateway
166...封包資料網路(PDN)閘道166. . . Packet Data Network (PDN) gateway
172...存取服務網路(ASN)閘道172. . . Access Service Network (ASN) Gateway
174...行動IP本地代理(MIP-HA)174. . . Mobile IP Local Agent (MIP-HA)
176...認證、授權、記賬(AAA)伺服器176. . . Authentication, Authorization, Accounting (AAA) Server
178...閘道178. . . Gateway
從以下描述中可以更詳細地理解本發明,這些描述是以實例方式給出的,並且可以結合所附圖式加以理解,其中:
第1圖是影像中的像素的圖式;
第2A-2C圖分別是中心、行和列內插點的圖式;
第3圖是針對vid01_2X序列的基於邊緣的內插與五分接頭(tap)濾波器方法之間的訊框等級PSNR比較;以及
第4A圖是使用兩個分離的濾波器的基於邊緣的內插和自適應銳化系統的方塊圖;
第4B圖是使用兩個分離的濾波器的基於邊緣的內插和自適應銳化系統的更詳細的方塊圖;
第5圖是內插過程的流程圖;
第6圖是銳化過程的流程圖;
第7圖是聯合的基於邊緣的內插和自適應銳化濾波器的方塊圖;
第8圖是帶旗標像素的像素圖;
第9圖是中心、列和行像素的像素圖;
第10圖示出了向其應用濾波器係數的像素的跨度(span);
第11圖是組合的濾波方法的一個實施方式的流程圖;
第12圖是示出了為(a)中心、(b)列和(c)行像素的鄰近像素分配銳化參數的像素圖;
第13-16圖是識別用於導出邊界框中的被內插像素的原始像素的像素圖;
第17圖識別用於銳化的原始像素和所估計的被內插的像素;
第18A圖是在其中可以實施一個或多個所揭露的實施方式的示例性通信系統的系統圖;
第18B圖是可以在第18A圖所示的通信系統中使用的示例性無線傳輸/接收單元(WTRU)的系統圖;以及
第18C-18E圖是可以在第18A圖所示的通信系統中使用的示例性無線電存取網路和示例性核心網路的系統圖。The invention may be understood in more detail from the following description, which is given by way of example, and
Figure 1 is a diagram of pixels in an image;
Figure 2A-2C is a diagram of the interpolation points of the center, row and column;
Figure 3 is a frame-level PSNR comparison between the edge-based interpolation and the five-tap filter method for the vid01_2X sequence; and Figure 4A is an edge-based interpolation using two separate filters. And a block diagram of the adaptive sharpening system;
Figure 4B is a more detailed block diagram of an edge-based interpolation and adaptive sharpening system using two separate filters;
Figure 5 is a flow chart of the interpolation process;
Figure 6 is a flow chart of the sharpening process;
Figure 7 is a block diagram of a joint edge-based interpolation and adaptive sharpening filter;
Figure 8 is a pixmap of the flagged pixel;
Figure 9 is a pixmap of the center, column and row pixels;
Figure 10 shows the span of pixels to which filter coefficients are applied;
Figure 11 is a flow diagram of one embodiment of a combined filtering method;
Figure 12 is a pixmap showing the allocation of sharpening parameters for adjacent pixels of (a) center, (b) column, and (c) row of pixels;
Figures 13-16 are pixmaps identifying the original pixels used to derive the interpolated pixels in the bounding box;
Figure 17 identifies the original pixels used for sharpening and the estimated interpolated pixels;
Figure 18A is a system diagram of an exemplary communication system in which one or more disclosed embodiments may be implemented;
Figure 18B is a system diagram of an exemplary wireless transmit/receive unit (WTRU) that can be used in the communication system shown in Figure 18A; and Figure 18C-18E is a diagram that can be used in the communication system shown in Figure 18A A system diagram of an exemplary radio access network and an exemplary core network.
Claims (31)
確定與一內插點相關聯的一邊緣特性,該邊緣特性具有一邊緣大小和一邊緣角;
回應於該邊緣角來選擇一內插濾波器;以及
使用所選擇的內插濾波器來確定該內插點處的一像素值。A method for interpolating an image, the method comprising:
Determining an edge characteristic associated with an interpolated point, the edge characteristic having an edge size and an edge angle;
Selecting an interpolation filter in response to the edge angle; and using the selected interpolation filter to determine a pixel value at the interpolation point.
獲取多個邊緣特性,每個該邊緣特性與多個內插點中的各自的一個內插點相關聯,每個該邊緣特性具有一邊緣大小和一邊緣角;
回應於各自的邊緣角,為該多個內插點中的每個內插點選擇一內插濾波器;
使用相應的所選擇的內插濾波器來確定該多個內插點中的每個內插點的一像素值;以及
輸出該像素值。A computer readable medium for storing instructions that, when executed by a processor, cause the processor to:
Obtaining a plurality of edge characteristics, each of the edge characteristics being associated with a respective one of the plurality of interpolation points, each edge characteristic having an edge size and an edge angle;
Selecting an interpolation filter for each of the plurality of interpolation points in response to respective edge angles;
Determining a pixel value of each of the plurality of interpolation points using a corresponding selected interpolation filter; and outputting the pixel value.
一邊緣特性計算器,被配置用於為多個內插點中的每個內插點確定一邊緣特性;
一內插濾波器選擇器,被配置用於對該邊緣特性進行操作並回應性地產生針對該多個內插點中的每個內插點的一內插濾波器識別符;以及
一內插濾波器電路,被配置用於回應於該內插濾波器識別符而應用多個內插濾波器中的一個內插濾波器、並輸出針對該多個內插點的多個被內插的值。An interpolation device, the interpolation device comprising:
An edge characteristic calculator configured to determine an edge characteristic for each of the plurality of interpolation points;
An interpolation filter selector configured to operate the edge characteristic and responsively generate an interpolation filter identifier for each of the plurality of interpolation points; and an interpolation a filter circuit configured to apply one of the plurality of interpolation filters in response to the interpolation filter identifier and output a plurality of interpolated values for the plurality of interpolation points .
回應於一梯度資料而對一影像訊框中的一像素進行內插;以及
回應於該梯度資料,對該影像訊框中的像素進行銳化。A method comprising:
Interpolating a pixel in an image frame in response to a gradient data; and in response to the gradient data, sharpening pixels in the image frame.
一內插模組,被配置用於回應於一梯度資料對影像訊框中的一像素進行內插;以及
一銳化濾波器模組,被配置用於回應於該梯度資料來對該影像訊框中的像素進行銳化。A device comprising:
An interpolation module configured to interpolate a pixel in the image frame in response to a gradient data; and a sharpening filter module configured to respond to the gradient data to the image The pixels in the box are sharpened.
確定用於要被內插的一影像像素的一梯度資料;
根據該梯度資料,以使用一鄰近像素來確定一被內插的像素值;
基於梯度資料來識別要被銳化的影像像素;以及
使用至少部分地基於該梯度資料所確定的一濾波器強度來對影像像素進行銳化。A method comprising:
Determining a gradient data for an image pixel to be interpolated;
Determining an interpolated pixel value using a neighboring pixel based on the gradient data;
The image pixels to be sharpened are identified based on the gradient data; and the image pixels are sharpened using a filter intensity determined based at least in part on the gradient data.
確定用於要被內插的一影像像素的一梯度資料;
選擇性地銳化一鄰近的原始像素;
回應於是一中心、一列或一行像素的像素類別,選擇性地識別還要被內插的一鄰近像素;以及
根據一聯合銳化和內插濾波器以使用鄰近像素來確定一被內插和銳化的像素值。A method comprising:
Determining a gradient data for an image pixel to be interpolated;
Selectively sharpening a neighboring original pixel;
Responding to a pixel class of a center, a column or a row of pixels, selectively identifying a neighboring pixel to be interpolated; and using a joint sharpening and interpolation filter to determine an interpolated and sharp using neighboring pixels The pixel value.
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US201161442069P | 2011-02-11 | 2011-02-11 | |
| US201161535353P | 2011-09-15 | 2011-09-15 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| TW201301199A true TW201301199A (en) | 2013-01-01 |
Family
ID=45689053
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| TW101104396A TW201301199A (en) | 2011-02-11 | 2012-02-10 | Edge-based video interpolation for video and image upsampling |
Country Status (2)
| Country | Link |
|---|---|
| TW (1) | TW201301199A (en) |
| WO (1) | WO2012109528A1 (en) |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US9165350B2 (en) | 2013-09-13 | 2015-10-20 | Novatek Microelectronics Corp. | Image sharpening method and image processing device |
| TWI511088B (en) * | 2014-07-25 | 2015-12-01 | Altek Autotronics Corp | Method of generating azimuth images |
| US9548043B2 (en) | 2013-07-24 | 2017-01-17 | Novatek Microelectronics Corp. | Pixel value interpolation method and system |
| TWI720513B (en) * | 2019-06-14 | 2021-03-01 | 元智大學 | Image enlargement method |
Families Citing this family (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| KR102272108B1 (en) | 2015-02-27 | 2021-07-05 | 삼성전자주식회사 | Image processing apparatus and method |
| US10007970B2 (en) | 2015-05-15 | 2018-06-26 | Samsung Electronics Co., Ltd. | Image up-sampling with relative edge growth rate priors |
| EP3348058B1 (en) | 2015-09-25 | 2021-09-15 | Huawei Technologies Co., Ltd. | Adaptive sharpening filter for predictive coding |
| EP3354026B1 (en) * | 2015-09-25 | 2021-09-15 | Huawei Technologies Co., Ltd. | Apparatus and method for video motion compensation |
| CN108141602B (en) | 2015-09-25 | 2021-02-23 | 华为技术有限公司 | Video codec, encoding and decoding method, and computer-readable storage medium |
| CN107925772B (en) | 2015-09-25 | 2020-04-14 | 华为技术有限公司 | Apparatus and method for video motion compensation using selectable interpolation filters |
| KR102143736B1 (en) | 2015-09-25 | 2020-08-12 | 후아웨이 테크놀러지 컴퍼니 리미티드 | Video motion compensation apparatus and method |
| EP3503018B1 (en) * | 2017-12-22 | 2025-03-05 | Vestel Elektronik Sanayi ve Ticaret A.S. | Adaptive local contrast enhancement zone for scaled videos |
| CN117308791B (en) * | 2023-09-25 | 2025-07-22 | 福建农林大学 | Structure displacement measurement method and system based on video frame insertion and computer vision |
Family Cites Families (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP4053021B2 (en) * | 2004-05-12 | 2008-02-27 | 三洋電機株式会社 | Image enlargement apparatus and program |
-
2012
- 2012-02-10 TW TW101104396A patent/TW201301199A/en unknown
- 2012-02-10 WO PCT/US2012/024630 patent/WO2012109528A1/en active Application Filing
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US9548043B2 (en) | 2013-07-24 | 2017-01-17 | Novatek Microelectronics Corp. | Pixel value interpolation method and system |
| US9165350B2 (en) | 2013-09-13 | 2015-10-20 | Novatek Microelectronics Corp. | Image sharpening method and image processing device |
| TWI511088B (en) * | 2014-07-25 | 2015-12-01 | Altek Autotronics Corp | Method of generating azimuth images |
| TWI720513B (en) * | 2019-06-14 | 2021-03-01 | 元智大學 | Image enlargement method |
Also Published As
| Publication number | Publication date |
|---|---|
| WO2012109528A1 (en) | 2012-08-16 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| TW201301199A (en) | Edge-based video interpolation for video and image upsampling | |
| JP6964567B2 (en) | Sampling grid information for spatial layers in multi-layer video coding | |
| JP6141988B2 (en) | Adaptive upsampling of multi-layer video coding | |
| US9626745B2 (en) | Temporal multi-band noise reduction | |
| JP6175505B2 (en) | Cross-plane filtering for chroma signal enhancement in video coding | |
| US9251572B2 (en) | System and method of correcting image artifacts | |
| US9177363B1 (en) | Method and image processing apparatus for image visibility restoration | |
| US10044913B2 (en) | Temporal filter for denoising a high dynamic range video | |
| US10045050B2 (en) | Perceptual preprocessing filter for viewing-conditions-aware video coding | |
| EP3649780A1 (en) | Motion-compensation prediction based on bi-directional optical flow | |
| US20150304685A1 (en) | Perceptual preprocessing filter for viewing-conditions-aware video coding | |
| US20120081553A1 (en) | Spatial filtering for image signal processing | |
| TW200926763A (en) | Converting video and image signal bit depths | |
| WO2017071499A1 (en) | Image sharpening method and mobile terminal | |
| US9928575B2 (en) | Method and system for processing a multi-channel image | |
| KR101858212B1 (en) | Inverse telecine filter | |
| US20250252534A1 (en) | Film grain synthesis using multiple correlated patterns | |
| JP2012238099A (en) | Image processing device, image processing method, and program | |
| WO2025061732A1 (en) | Histogram normalization associated with decoder side intra mode derivation | |
| WO2025073738A1 (en) | Cross component filter coefficients in bif | |
| WO2025073660A1 (en) | Cross component scaling factor in bif | |
| CN115829871A (en) | Image denoising method and device, chip and module equipment | |
| WO2024133624A1 (en) | Local illumination compensation with extended models | |
| WO2024133767A1 (en) | Motion compensation for video blocks | |
| WO2023057445A1 (en) | Video sample filtering based on depth or motion information |