Loading...
Welcome to Journal of Graphics share: 

Table of Contents

    For Selected: Toggle Thumbnails
    Cover
    Cover of issue 4, 2022
    2022, 43(4): 0. 
    Abstract ( 129 )   PDF (1594KB) ( 157 )  
    Related Articles | Metrics
    Contents
    Table of Contents for Issue 4, 2022
    2022, 43(4): 1-1. 
    Abstract ( 77 )   PDF (219KB) ( 90 )  
    Related Articles | Metrics
    Image Processing and Computer Vision
    Weed detection in vegetable field based on improved YOLOv4 and image processing
    DONG Hui, CHEN Xin-kai, SUN Hao, YAO Li-gang
    2022, 43(4): 559-569.  DOI: 10.11996/JG.j.2095-302X.2022040559
    Abstract ( 655 )   PDF (6215KB) ( 402 )  
    To address the problems of low efficiency, poor accuracy, and insufficient robustness of detection methods due to the variety and complex distribution of weeds in the field, the seven kinds of common vegetables and field weeds in the seedling stage in the seedling field were taken as the research objects, the weed detection was reversely converted into crop detection, and a weed detection algorithm in vegetable seedling fields based on optimized YOLOv4 and image processing was proposed. Based on the YOLOv4 object detection algorithm, the backbone feature extraction network was embedded in SA module to enhance the feature extraction capability, the Transformer module was introduced to construct the long-distance global semantic information of the feature map, and the detection head and loss function were improved to increase the detection and positioning accuracy. The improved model’s average recognition time for a single image was 0.261 s, and the average recognition accuracy rate was 97.49%. Under the same training samples and system environment settings, the improved method was compared with the mainstream target detection algorithms Faster RCNN, SSD, and YOLOv4. The results show that the improved YOLOv4 model is of evident advantages in the identification of various vegetables in the seedling stage. The improved YOLOv4 target detection algorithm was used to detect crops: the vegetation outside the crop area is weeds, and the excess-green feature was combined with the OTSU threshold segmentation algorithm to obtain the weed foreground. Finally, the connected component of the weed foreground was marked to output the weed centroid coordinates and the position of the detection frame. In doing so, weed can be effectively detected in vegetable seedling fields.
    Related Articles | Metrics
    An improved low-dose CT image enhancement network based on CycleGAN
    LIAO Shi-min, LIU Yang-chuan, ZHU Ye-chen, WANG Yan-ling, GAO Xin
    2022, 43(4): 570-578.  DOI: 10.11996/JG.j.2095-302X.2022040570
    Abstract ( 362 )   PDF (2313KB) ( 239 )  
    Low-dose CT is an effective and relatively safe screening method for thoracic and abdominal diseases, but the artifacts and noise in the image will seriously affect the doctor’s diagnosis. Network training in image enhancement methods based on deep learning mostly relies on paired data that is pixel-level matched low-dose and conventional-dose CT images at the same site of the same patient. An improved low-dose CT image enhancement network based on the cycle-consistent generative adversarial network (CycleGAN) was proposed for unpaired data. A shallow feature pre-extraction module was added in front of the generator to enhance the capability to extract CT images features. In addition, the depthwise separable convolution was used to replace some common convolutions in the generator to decrease network parameters and reduce GPU memory usage. In the proposed network, a total of 3 275 two-dimensional low-dose CT slices and a total of 2 790 two-dimensional unpaired conventional-dose CT slices were used for training, and a total of 1 716 two-dimensional low-dose CT slices were employed for testing. The results show that the averaged perception-based image quality evaluator (PIQE) of CT images generated by the network is 45.53, which is 8.3% lower than that of CycleGAN, 31.9% lower than that of Block-Matching and 3D filtering (BM3D), and 20.9% lower than that of unsupervised image translation network (UNIT). Moreover, the proposed network can produce better subjective visual effects in terms of structural detail preservation, noise and artifact suppression. This shows that the network is a low-dose CT image enhancement method with potential clinical prospects.
    Related Articles | Metrics
    Research and application of wild mushrooms classification based on multi-scale features to realize hyperparameter evolution
    ZHANG Dun, HUANG Zhi-kai, WANG Huan, WU Yi-peng, WANG Ying, ZOU Jia-hao
    2022, 43(4): 580-589.  DOI: 10.11996/JG.j.2095-302X.2022040580
    Abstract ( 157 )   PDF (5924KB) ( 141 )  
    In China, there are frequent poisoning events caused by ingestion of inedible wild mushrooms every summer, especially in Southwest China, such as Yunnan. This is due to the slight differences in inter-class characteristics of wild mushrooms and the complex image backgrounds in actual scenarios, making it difficult to distinguish only by naked eyes. At present, although there are many methods to classify wild mushrooms, and the most reliable way is molecular identification, the relevant techniques are time-consuming and require a high threshold, so they are not suitable for real-time classification and detection. To solve this problem, an approach based on deep learning was proposed. This approach employed the attention mechanism convolution block attention module (CBAM), was combined with multi-scale fusion, and added the anchor layer. The hyperparameter evolution idea was adopted to adjust the hyperparameter during the model training, so as to improve the recognition accuracy. Compared with standard target detection networks, such as SSD, Faster Rcnn, and Yolo series, the proposed model can classify and detect wild mushrooms more accurately. Compared with the original Yolov5, after the proposed model was improved, Map was improved by 3.7% and reached 93.2%, precision by 1.3%, Recall by 1.0%, and model detection speed by 2.3%. Compared with SSD, Map was improved by 14.3%. Finally, the model was simplified and deployed on Android devices to increase its practicability, thus solving the current problem of poisoning caused by eating inedible wild mushrooms because of the difficulty of identification.
    Related Articles | Metrics
    Multi category mask wearing detection based on dynamic weighted category balance loss
    CHEN Zhao-jun, CHU Jun, ZENG Lun-jie
    2022, 43(4): 590-598.  DOI: 10.11996/JG.j.2095-302X.2022040590
    Abstract ( 125 )   PDF (1907KB) ( 123 )  
     Mask wearing in public has become an important measure to control the spread of Coronavirus Disease 2019 (COVID-19). With the prolonged development of the COVID-19 epidemic, the public’s awareness of self-protection has been gradually declining, leading to the increasing tendency of wearing masks incorrectly in public. The existing mask wearing detection methods usually only detect whether the mask is worn, without the detection of non-standard mask wearing scenarios, which is likely to cause cross infection. The current mask datasets lack the image data of non-standard mask wearing. To solve the above problems, on the basis of the existing mask datasets, more non-standard mask wearing images were collected through the Internet and offline, and the Mosaic data enhancement algorithm was improved to expand the data according to the features of face images in the cases of wearing masks. The improved Mosaic data enhancement algorithm could improve the mean average precision (mAP) of the benchmark network YOLOv4 by 2.08%. To address the problem of category imbalance in the dataset after data enhancement, the dynamic weighted balance loss function was proposed. Based on the weight binary cross entropy loss function, the reciprocal of the number of effective samples served as the auxiliary category weight, and dynamic adjustment was performed in each batch under training, thus solving the problems of weak stability, precision oscillation, and unsatisfactory effect when the re-weighting method was directly put to use. The experiment showed that mAP of the improved model reached 91.25%, and the average precision (AP) of non-standard mask wearing reached 91.69%. Compared with such single-stage methods as RetinaNet, Centernet, and Effcientdet, and such two-stage methods as YOLOv3-MobileNetV2 and YOLOv4-MobileNetV2, the improved algorithm exhibits higher detection accuracy and speed.
    Related Articles | Metrics
    A scattered point cloud simplification algorithm based on FPFH feature extraction
    LI Hai-peng, XU Dan, FU Yu-ting, LIU Yan-an, ZHANG Ting-ting
    2022, 43(4): 599-607.  DOI: 10.11996/JG.j.2095-302X.2022040599
    Abstract ( 181 )   PDF (4249KB) ( 112 )  
    To address the large amount of redundant data in the original point cloud, a point cloud simplification algorithm based on fast point feature histograms (FPFH) feature extraction was proposed, effectively taking into account the retention of feature information and overall integrity. Firstly, the algorithm sought and retained the boundary points of the original model. Then, the FPFH value of non-boundary points were calculated, thus producing the feature value of the point cloud. After sorting the feature values, the non-boundary points were divided into the feature region and the non-feature region, retaining the points in the feature region. Finally, the non-feature region was divided into k sub-intervals, and the improved farthest point sampling algorithm was employed to sample each sub-interval. The proposed algorithm was compared with the farthest point sampling algorithm, non-uniform grid method, k-means algorithm, and adaptive curvature entropy algorithm, and the simplified point cloud was evaluated by the standardized information entropy evaluation method. Experimental results show that the proposed algorithm outperforms other simplification algorithms. In addition, the visualization results indicate that the proposed algorithm can not only ensure the integrity of the simplified model but also retain most of the feature information of the point cloud.
    Related Articles | Metrics
    A computer vision based structural damage identification method for temporary structure during construction
    LIANG Zhen-yu, HUA Jia-hao, CHEN Hao-long, DENG Yi-chuan
    2022, 43(4): 608-615.  DOI: 10.11996/JG.j.2095-302X.2022040608
    Abstract ( 127 )   PDF (2999KB) ( 103 )  
    Temporary structure is the main risk source of construction site accidents. Previous vibration-based detection methods mainly focus on setting accelerometers on some pre-defined critical areas. However, due to the factors such as nonstandard component erection and uncertainty of the construction site for the temporary structure, the critical areas of the monitoring obtained from the analysis may vary dramatically from the reality. Therefore, this paper proposed a structural damage identification method for temporary structure based on phased-based Eulerian video magnification algorithm, making full use of the advantages of global coverage and efficient monitoring of computer vision technology. The digital image of temporary structure vibration collected by digital camera was firstly processed by phased-based Eulerian video magnification to acquire motion-magnified image sequence in the particular frequency bands. Then, the canny edge detector was employed to identify the edges in the image sequence and eliminate the noise resulting from the magnification. The edges in the image sequence were utilized to acquire time-history data of temporary structure displacement based on the geometry centroid, from which resonant frequencies could be obtained after Fourier transformation, and finally the damage states were identified based on the pre-established damage dynamic fingerprint
    database. The applicability of the proposed method was discussed in the context of the frame scaffold experiments with 10 kinds of damage states. By comparing the results between camera measurement and accelerometer measurement, the proposed method can yield satisfactory performance with an average error of 0.95%, fulfilling the accuracy requirements of damage identification.
    Related Articles | Metrics
    Point cloud classification and segmentation based on ring query and channel attention
    LIU Yu-zhen, LI Nan, TAO Zhi-yong
    2022, 43(4): 616-623.  DOI: 10.11996/JG.j.2095-302X.2022040616
    Abstract ( 77 )   PDF (1057KB) ( 71 )  
    Feature processing of point cloud data is a key component of 3D object recognition technology in robotics, autopilot, and other fields. In order to address the problems of repeated extractions of local feature information of point cloud and lack of recognition of the whole geometric structure of point cloud object, a point cloud classification and segmentation network based on ring query and channel attention was proposed. First the single-layer ring query was combined with the feature channel attention mechanism to reduce local information redundancy and strengthen local features. Then the high response points of the edges and corners of the object were identified by calculating the normal changes, and the normal features were added to the global feature representation, thereby strengthening the recognition of the whole geometric structure of the object. Compared with many point-cloud networks on ModelNet40 and ShapeNet Part datasets, the experimental results show that the network not only has higher accuracy for point cloud classification and segmentation, but also outperforms other methods in training time and memory consumption. In addition, the network is strongly robust for the number of different input point clouds. Therefore, the proposed network is an effective and feasible network for point cloud classification and segmentation.
    Related Articles | Metrics
    Zero-shot image classification based on random propagation graph convolution model
    LU Nan-nan, LIU Yi-xiong, QIU Ming-kai
    2022, 43(4): 624-631.  DOI: 10.11996/JG.j.2095-302X.2022040624
    Abstract ( 104 )   PDF (1744KB) ( 76 )  
    Zero-shot image classification aims to recognize new categories, namely, unseen categories that do not appear during training. Therefore, auxiliary information is needed to model the relationship between unseen and seen categories. With the aid of knowledge graph, zero-shot classification models based on graph convolution network (GCN) can explicitly express the relationship between categories, but GCN is susceptible to over-smoothing, resulting in the degradation of model performance. To address this problem, a zero-shot classification model based on random propagation graph convolution was proposed. In this model, the raw features were processed by random propagation mechanism to achieve feature perturbation and data augmentation. The generated knowledge graph based on category hierarchy could model the semantic relationship between categories, where graph nodes stand for categories and graph edges stand for relationships. Then the GCN was constructed to train the processed features, and the classifier parameters containing unseen categories, which were the output of nodes, could achieve zero-shot classification. Experimental results show that the model can significantly decrease time consumption, and improve accuracy and generalization performance.
    Related Articles | Metrics
    A method for obtaining the complete point cloud of reinforcement skeletons based on a structured light camera
    LIU Shi-long, MA Zhi-liang
    2022, 43(4): 633-640.  DOI: 10.11996/JG.j.2095-302X.2022040633
    Abstract ( 104 )   PDF (1663KB) ( 89 )  
    In order to obtain the high-precision complete point cloud of a reinforcement skeleton required for the automatic quality inspection, an algorithm was proposed for obtaining the complete point cloud of reinforcement skeletons based on a structured light camera. Firstly, 3D reconstruction was carried out for multiple reinforcement skeleton images collected using a structured light camera, thus obtaining the dimensionless poses of the structured light camera. Secondly, the dimensional poses were obtained according to the dimensionless poses. Then, the precise transformation matrix between these dimensional poses was calculated. Next, based on these dimensional poses and the precise transformation matrix between them, graph optimization was employed to optimize these dimensional poses to obtain those with high precision. Finally, point clouds obtained using the structured light camera were aligned based on the dimensional poses with high precision, which can generate the complete point cloud of the reinforcement skeleton. The experimental results show that it would take the proposed algorithm about 10 minutes to obtain the complete point cloud of the reinforcement skeleton of a practical precast concrete component, and the error of the point cloud is around 5 mm. It is concluded that the proposed algorithm can quickly obtain the complete point cloud of the reinforcement skeleton, with high accuracy, which can meet the requirement of the automatic quality inspection of the reinforcement skeleton.
    Related Articles | Metrics
    Unsupervised emotion recognition of Yunnan Heavy Color Paintings based on domain adaptation
    PENG Guo-qin, ZHANG Hao, XU Dan
    2022, 43(4): 641-650.  DOI: 10.11996/JG.j.2095-302X.2022040641
    Abstract ( 93 )   PDF (10685KB) ( 122 )  
    Thanks to the large-scale labeled datasets available, deep learning has made a great breakthrough in computer vision. However, due to the ambiguity of emotion semantics, it is hard to annotate the emotional labels for images. Thus, only a few small-scale image emotion datasets are open and available, restricting the performance of image emotion analysis based on deep learning. The semantics of emotions have unique characteristics, such as order and polarity, but few studies have paid attention to these essential characteristics in image emotion analysis. Thus, in the paper, based on domain adaptation, considering the essential characteristics of emotion semantics, that is, the ordered and grouped polarity, we proposed to measure emotion semantic differences through earth mover’s distance (EMD). The goal is to better transfer the trained model with labeled emotion dataset to unlabeled emotion dataset and complete the unsupervised image emotion analysis. The Yunnan Heavy Color Paintings Emotion dataset was created in this paper, and was applied to our proposed method. The experimental results demonstrate that the proposed method can effectively align the emotional semantics between the source domain and the target domain, realizing the unsupervised automatic annotation of emotion dataset, thus expanding the size of the image emotion dataset.
    Related Articles | Metrics
    Multi-frame compressed video enhancement based on spatio-temporal fusion
    MA Yan-bo, LI Lin, CHEN Yuan, ZHAO Yang, HU Rui
    2022, 43(4): 651-658.  DOI: 10.11996/JG.j.2095-302X.2022040651
    Abstract ( 98 )   PDF (3765KB) ( 74 )  
    In order to reduce the storage and transmission cost of video, lossy compression is in frequent use, which however would incur various types of artifacts in the video and affect users’ subjective visual experience. The single frame method cannot be directly applied to video processing, because they independently process each video frame, limiting the use of spatial information and causing limited effectiveness. Inter-frame alignment or temporal structure was often adopted in multi-frame methods to enhance the reconstruction results by utilizing the temporal information, but there remains much room for improvement in alignment performance. To solve the above problems, a multi-frame spatio-temporal compression artifact removal method was proposed to achieve better alignment fusion through the alignment fusion design. This method efficiently utilized the multi-frame spatio-temporal information to reconstruct high quality videos. The experimental results show that the proposed method outperforms other comparative methods on a number of test compressed videos with different resolutions (HM16.5 under low delay P), and that it can achieve an average improvement of 0.13 dB on the objective index peak signal to noise ratio (PSNR) compared with the state-of-the-art multi-frame method STDF. Meanwhile, the proposed method can yield promising results in subjective comparisons, reconstructing a clearer picture with a good effect of compression artifact removal.
    Related Articles | Metrics
    Grayscale watermarking algorithm via BEMD and texture complexity
    ZHAO Hui-chao, HU Kun, WANG Xiao-chao
    2022, 43(4): 659-666.  DOI: 10.11996/JG.j.2095-302X.2022040659
    Abstract ( 123 )   PDF (5489KB) ( 67 )  
    This paper presented a grayscale watermarking algorithm combining texture complexity and bi-dimensional empirical mode decomposition (BEMD). Firstly, the watermark image after Arnold transform was decomposed by BEMD to obtain the intrinsic modal function (IMF) and the residual information of different scales. Then, the texture complexity of the host image after segmentation was calculated, and the region of high texture complexity was selected as the embedding location. In order to better fuse the IMF of the watermark image with the host image, BEMD decomposition was performed on the host image under the same sieving conditions as those for the watermark image. Finally, the watermark information was repeatedly embedded into the pre-selected location of the host image, and then the image embedded with watermarks was reconstructed by the IMF and the residual. Watermark extraction was the inverse process of watermark embedding. The experiments on host images at different texture levels revealed that the peak signal-to-noise ratios of images embedded with watermarks were all above 40 dB, and the normalized coefficients values of watermark extraction all exceeded 0.95 in the face of eight common attacks. Compared with the existing algorithms, the proposed algorithm performs well in large scale shearing, noise attack, image filtering, and JPEG compression attack, superior to the algorithms under comparison.
    Related Articles | Metrics
    Computer Graphics and Virtual Reality
    WebAR garbage classification information visualization method based on VD-MobileNet network
    LIU Nan-shan, PEI Yun-qiang, JIANG Hao, HAN Yong-guo, WU Ya-dong, WANG Fu-pan, YI Si-heng
    2022, 43(4): 667-676.  DOI: 10.11996/JG.j.2095-302X.2022040667
    Abstract ( 132 )   PDF (2845KB) ( 104 )  
    With the accelerated implementation of the garbage classification regulation in China, many applications for garbage classification based on virtual/augmented reality technologies have sprung up. Under the influence of the identification equipment platform and residents’ habits of using applications, there remain a number of shortcomings in convenience and practicability for this kind of application. A waste classification application scheme was proposed based on a lightweight neural network combined with mobile augmented reality and visualization technology. Firstly, the variable expansion convolution VD-MobileNet model method was proposed for garbage classification based on deep learning, which can solve the problems of limited computing capacity and a huge network of mobile devices. The receptive field was increased by introducing dilated convolution in the MobileNet model. The characteristic information of garbage could be expanded to enhance classification accuracy, and LeakyReLU activation function was
    introduced to optimize the expression ability of the network. Secondly, the model was equipped with the WebAR technology, and a lightweight garbage classification information visualization system was designed for mobile devices. This system could operate cross different platforms, realize the diversified visual presentation of classified information, and enable flexible interactions. Experiments and evaluations show that the VD-MobileNet model could achieve excellent classification in the garbage classification data set and can effectively reduce the amount of calculation with constant parameters. In addition, the WebAR application system designed based on the model can provide users with reasonable and effective assistance in garbage disposal.
    Related Articles | Metrics
    Sea surface temperature prediction algorithm combined with residual spatial-temporal attention mechanism
    HE Qi, LI Wen-long, SONG Wei, DU Yan-ling, HUANG Dong-mei, GENG Li-jia
    2022, 43(4): 677-684.  DOI: 10.11996/JG.j.2095-302X.2022040677
    Abstract ( 145 )   PDF (833KB) ( 112 )  
    Sea surface temperature (SST) is closely related to global climate change, ocean disasters, and ocean ecosystems, so the accurate prediction of SST is an important topic. The existing regional SST prediction methods treat the time series of SST data as a series of matrixes, each corresponding to the regional SST at a particular time. The spatial and temporal features are extracted from the matrix series for later SST prediction. However, the existing SST prediction methods fail to fully consider the imbalanced influence of temporal and spatial features on the SST, leading to the neglection of some key information and limiting the improvement of prediction accuracy. To address this problem, we proposed a regional SST prediction method (CRA-ConvLSTM) combining temporal attention mechanism and spatial attention mechanism. This enabled the model to dynamically assign different influence weights to the temporal features at different times and spatial features at different locations, thereby improving the accuracy of SST prediction. Specifically, the input regional SST time series was first encoded into multi-layer feature vectors by a convolutional neural network (CNN), and local features were extracted. Then, the residual time attention module was constructed to learn the attention weight at different moments adaptively, and the key features of the time dimension were extracted. The residual spatial attention module was designed to extract the key features of different points in the region in terms of the spatial dimension. In addition, the attention mechanism combined with the residual structure can avoid performance degradation caused by information reduction in the network. Experimental results show that the proposed model could achieve 0.19 and 99.43% respectively in terms of the root mean square error (RMSE) and prediction accuracy (PACC), which is superior to other methods and effectively improves the prediction accuracy of SST.
    Related Articles | Metrics
    A visual analysis approach for domain literature data based on word representation model
    ZHANG Qing-hui, CHEN Yi, WU Cai-xia
    2022, 43(4): 685-694.  DOI: 10.11996/JG.j.2095-302X.2022040685
    Abstract ( 107 )   PDF (4313KB) ( 74 )  
    With the development of science and technology, scientific literature is mounting to an increasingly large scale. How to quickly and accurately seek the research topics, influential scholars, and high-level papers in a specific domain from the vast amount of publications remains an enormous challenge. The visual analysis method for domain literature data based on word representation model employed word2vec to recommend domain-related keywords by the similarity between word vectors, and filters the domain-related papers according to these keywords. Then it utilized the BERTopic model to extract topics from the abstracts of domain papers. Next, the values for paper impact were calculated using PageRank, and the values for author influence were calculated using Author-Rank, the author impact evaluation method, taking into account the order of authorship, the number of publications, and the impact of papers. Finally, the multi-view collaborative and interactive visualization approach could help researchers gain a quick understanding and analysis of specific areas from multiple perspectives, such as topics word frequency, topics evolution, literature impact, citation relationships, and author impact. The method can be applied to literature data analysis in the field of “food safety”, and the results and user tests can validate this method.
    Related Articles | Metrics
    Human pose estimation and similarity calculation for Tai Chi learning
    CAI Xing-quan, HUO Yu-qing, LI Fa-jian, SUN Hai-yan
    2022, 43(4): 695-706.  DOI: 10.11996/JG.j.2095-302X.2022040695
    Abstract ( 271 )   PDF (11185KB) ( 194 )  
    To address the current problems of poor natural interactivity and lack of learning feedback in the case of online Tai Chi learning, this paper proposed a method of human pose estimation and similarity calculation for Tai Chi learning. First, the proposed method extracted the key-frame images from the Tai Chi video using an inter-frame difference method. Second, our method employed the stacked hourglass network model to perform two-dimensional joint-point detection on the key-frame images. Third, a long short-term memory (LSTM) network combined with the Sequence-to-Sequence network model was used to map the detected two-dimensional joint-point sequence from two-dimensional to three-dimensional, thus predicting the position coordinates of the three-dimensional joint-points. Finally, the two-dimensional and three-dimensional cosine similarities of the estimated human posture were calculated. Using this method, this paper designed and developed a Tai Chi learning and feedback application system with simple equipment and strong user experience, which was applied to real scenarios. This system could detect whether the overall movements of Tai Chi students and the movements of each body segment were standard, with feedback provided. Students could practice and improve non-standard movements based on the feedback, so as to achieve the purpose of improving the learning effect.
    Related Articles | Metrics
    Virtual light-based translucent material rendering in real-time
    CHEN Zhu-xin, YANG Qin-qi, CHEN Rui, ZHANG Yan-ci, LIU Yan-li, WU Zhi-hong
    2022, 43(4): 707-714.  DOI: 10.11996/JG.j.2095-302X.2022040707
    Abstract ( 153 )   PDF (2451KB) ( 101 )  
    Translucent material rendering is an essential part of the field of real-time rendering. Transmission rendering relies on the accurate calculation of transmission thickness and is often limited by the complexity of scene models and lighting. This paper proposed a virtual light-based method to calculate the transmission thickness in translucent material rendering. A virtual light was added to the scene, and the sorting algorithm was employed at the virtual light for the depth information of the scene. In calculating the transmission thickness from the real light source to the shading point, it was proposed to take samples on a straight line segment of the world space connecting the two points, calculate the proportion of sampled points inside the object to the total number, which was multiplied by the length of the straight line segment, thus obtaining the estimation. Moreover, in the case of multiple real light sources
    in the scene, the sampling-based method could reuse the scene depth information stored in the virtual light. The proposed method could effectively enhance the accuracy of transmission thickness calculation, and reduce the memory overhead caused by the increasing number of light sources. Experiments show that the proposed method can strike a good balance between efficiency, effectiveness, and memory overhead.
    Related Articles | Metrics
    Digital Design and Manufacture
    Research on microstructure region identification and path splicing method of abrasive cloth wheel polishing blade
    LIU Jia, ZHANG Jing-jing, YANG Sheng-qiang, QIAO Zhi-jie
    2022, 43(4): 715-720.  DOI: 10.11996/JG.j.2095-302X.2022040715
    Abstract ( 57 )   PDF (1740KB) ( 48 )  
    The blade profile is characteristic of abrupt curvature and needs to be processed in different areas. The precise identification of the microstructure area and the splicing of the polishing path are the key to improving the consistency of the blade surface quality. To address this problem, this paper proposed to identify the front and rear edge microstructure areas based on the tangent vector angle of the section line, and to identify the root transition arc microstructure area based on the normal vector registration of the section line. According to the matching of the maximum processing belt width and the polishing point, the transition arc microstructure area of the front, the rear edge, and the blade root was spliced with the blade pot and the blade back polishing path, respectively. Simulation and experimental results show that compared with traditional arc recognition methods, the proposed method can more effectively retain the contour information of the microstructure area. Compared with the polishing method without path splicing, the accuracy of the blade profile after polishing increased by 49.52%, the surface roughness by 57.31%, and the consistency of the processing quality by 7.15% and 11.55%. These results prove that the identification of the microstructure area and the path splicing can effectively improve the consistency of the blade processing quality.
    Related Articles | Metrics
    BIM/CIM
    Crack visualization management method based on computer vision and BIM
    XIONG Chen, CHEN Li-bin, LI Lin-ze, XU Zhen, ZHAO Yang-ping
    2022, 43(4): 721-728.  DOI: 10.11996/JG.j.2095-302X.2022040721
    Abstract ( 287 )   PDF (1249KB) ( 227 )  
    Continuous monitoring and management of structural surface cracks is important to structural safety. To achieve automated structural crack identification and management, a series of crack identification, vectorization, and visualization methods were proposed based on computer vision and building information modeling (BIM). Firstly, the raster images of crack skeleton were extracted from structure surface images based on a deep learning method. Secondly, an automated vectorization method for the raster images of crack skeleton was proposed to obtain the coordinates of key points of cracks. Finally, the automated modeling and visualization of cracks were realized using Dynamo programming on BIM platform. The proposed crack vectorization method can obtain the topological information of cracks and significantly reduce the amount of stored data, thus facilitating crack visualization. In addition, through the collision analysis of BIM components, to which components the cracks belonged to can be easily identified. The component information and the crack width information can be stored as attribute data of each crack. The proposed method can attain an automated crack vectorization and visualization, providing a useful reference for large-scale crack identification and management.
    Related Articles | Metrics
    Research on recognition method of overlapped characters in construction drawings based on adaptive scale edge feature
    WANG Zheng, DENG Xue-yuan
    2022, 43(4): 729-735.  DOI: 10.11996/JG.j.2095-302X.2022040729
    Abstract ( 119 )   PDF (943KB) ( 76 )  
    At present, the recognition technology of non-overlapped characters has been perfected, but it remains difficult to solve the recognition problem of common overlapped characters in scenarios such as the annotation of architectural engineering drawings, which hinders the breakthrough of automatic modeling technology based on 2D scanned drawings. To address the incapability of traditional character recognition methods to recognize overlapped characters, a new method was proposed for overlapped characters recognition in construction drawings based on adaptive scale edge features. Based on the spatial distribution characteristics of pixels, the overlapped character areas were preliminarily determined, and the adaptive scale edge features of characters were defined and extracted. The result combination of “position + content” was screened with the help of the bivariate matching probability function, and the global optimal principle was used instead of the absolute threshold as the identification standard. Finally, the correct recognition of overlapped characters was achieved. Different from the conventional idea of recognizing after repairing, the new method combined feature matching and interference filtering, character positioning and character recognition. The proposed method can solve the overlapping character recognition problem insolvable for mature commercial OCR such as Baidu, and the data experiment proves that this method is of high recognition accuracy.
    Related Articles | Metrics
    Industrial Design
    Research on interactive design of children’s sound and painting based on multimodal sensory experience
    LI Xiao-ying, YU Ya-ping
    2022, 43(4): 736-743.  DOI: 10.11996/JG.j.2095-302X.2022040736
    Abstract ( 223 )   PDF (3023KB) ( 156 )  
    In order to solve the problems of children’s limited thinking and insufficient interaction caused by the one-sided inculcation of children’s picture education and the traditional way of drawing, the audio-visual interactive system for children based on multi-modal sensory experience was constructed. Based on the characteristics of children’s sensory capacity, the system integrated the audio-visual sensory attributes and matched the design requirements of children’s audio-visual interaction according to the relationship between human and objects in interaction design, and constructed children’s audio-visual interactive behaviors. The visualization technology was applied to the establishment of the mapping of sound and picture and the audio-visual interactive platform for children. Through the innovated painting experience, children’s sound and painting creation were made richer and more diversified, and the boundary of their cognitive ability was expanded. The fuzzy Delphi method was utilized to collect and screen the index factors
    affecting children’s painting and to determine the index model of vocal painting experience. The fuzzy analytic hierarchy process (AHP) was employed to calculate the weight of each index element. The feasibility of the platform was verified by experiments regarding children users, which were a combination of interviews and fuzzy comprehensive evaluation. Children’s vocalization served as the medium of painting creation, which could change the traditional drawing method using hands, brain, and eyes, thereby optimizing their interactive painting experience and heighten their interest in painting creation.
    Related Articles | Metrics
    Evaluation and optimization of TMA interface supported eye movement technology
    DUAN Yan-hua, LIU Zi-jian, NING Duo
    2022, 43(4): 745-752.  DOI: 10.11996/JG.j.2095-302X.2022040745
    Abstract ( 106 )   PDF (2640KB) ( 79 )  
    Usability evaluation was performed for the collagen fiber thermal deformation analyzer TMA (temperature achine analyzer) user interface using eye-tracking technology, and an optimized design solution was proposed. Taking the existing collagen fiber thermal deformation analyzer TMA interface as a prototype, under experts’ guidance, four design elements were selected, and two user interfaces with different styles were redesigned. Eye movement experiments were performed, eye movement data was collected, and SPSS was employed to conduct data analysis. Combined with questionnaire surveys, the usability was evaluated in terms of background color and layout types of the user interfaces. The analyzer’s program background color and layout types could affect the experience of tested users. The T-shaped layout outperformed the upper and lower frame layout in all aspects. The TMA user interface of the collagen fiber thermal deformation analyzer was optimized by the eye tracking technology. It optimized the deficiencies of the experience-based designers who subjectively designed the original interface, and proposed the design principles for the program of experimental instruments. It could thus shed lights on the concepts for the interactive interface style of the experimental instrument and enhance the soundness and rationality of the design.
    Related Articles | Metrics
    Published as
    Published as 4, 2022
    2022, 43(4): 753-753. 
    Abstract ( 39 )   PDF (73970KB) ( 62 )  
    Related Articles | Metrics