Search Results for author: Hidetoshi Shimodaira

Found 27 papers, 12 papers with code

Block-Diagonal Orthogonal Relation and Matrix Entity for Knowledge Graph Embedding

1 code implementation11 Jan 2024 Yihua Zhu, Hidetoshi Shimodaira

The primary aim of Knowledge Graph embeddings (KGE) is to learn low-dimensional representations of entities and relations for predicting missing facts.

Knowledge Graph Embedding Knowledge Graph Embeddings +2

Axis Tour: Word Tour Determines the Order of Axes in ICA-transformed Embeddings

1 code implementation11 Jan 2024 Hiroaki Yamagiwa, Yusuke Takase, Hidetoshi Shimodaira

Inspired by Word Tour, a one-dimensional word embedding method, we aim to improve the clarity of the word embedding space by maximizing the semantic continuity of the axes.

Word Embeddings

Knowledge Sanitization of Large Language Models

1 code implementation21 Sep 2023 Yoichi Ishibashi, Hidetoshi Shimodaira

We explore a knowledge sanitization approach to mitigate the privacy concerns associated with large language models (LLMs).

Question Answering

3D Rotation and Translation for Hyperbolic Knowledge Graph Embedding

1 code implementation22 May 2023 Yihua Zhu, Hidetoshi Shimodaira

The main objective of Knowledge Graph (KG) embeddings is to learn low-dimensional representations of entities and relations, enabling the prediction of missing facts.

Knowledge Graph Embedding Relation +1

Discovering Universal Geometry in Embeddings with ICA

1 code implementation22 May 2023 Hiroaki Yamagiwa, Momose Oyama, Hidetoshi Shimodaira

This study utilizes Independent Component Analysis (ICA) to unveil a consistent semantic structure within embeddings of words or images.

Norm of Word Embedding Encodes Information Gain

no code implementations19 Dec 2022 Momose Oyama, Sho Yokoi, Hidetoshi Shimodaira

Distributed representations of words encode lexical semantic information, but what type of information is encoded and how?

Informativeness Keyword Extraction +2

Improving word mover's distance by leveraging self-attention matrix

1 code implementation11 Nov 2022 Hiroaki Yamagiwa, Sho Yokoi, Hidetoshi Shimodaira

The proposed method is based on the Fused Gromov-Wasserstein distance, which simultaneously considers the similarity of the word embedding and the SAM for calculating the optimal transport between two sentences.

Paraphrase Identification Semantic Similarity +2

Improving Nonparametric Classification via Local Radial Regression with an Application to Stock Prediction

no code implementations28 Dec 2021 Ruixing Cao, Akifumi Okuno, Kei Nakagawa, Hidetoshi Shimodaira

For correcting the asymptotic bias with fewer observations, this paper proposes a \emph{local radial regression (LRR)} and its logistic regression variant called \emph{local radial logistic regression~(LRLR)}, by combining the advantages of LPoR and MS-$k$-NN.

regression Stock Prediction

Revisiting Additive Compositionality: AND, OR and NOT Operations with Word Embeddings

no code implementations18 May 2021 Masahiro Naito, Sho Yokoi, Geewook Kim, Hidetoshi Shimodaira

(Q2) Ordinary additive compositionality can be seen as an AND operation of word meanings, but it is not well understood how other operations, such as OR and NOT, can be computed by the embeddings.

Word Embeddings

Extrapolation Towards Imaginary 0-Nearest Neighbour and Its Improved Convergence Rate

no code implementations NeurIPS 2020 Akifumi Okuno, Hidetoshi Shimodaira

The weights and the parameter $k \in \mathbb{N}$ regulate its bias-variance trade-off, and the trade-off implicitly affects the convergence rate of the excess risk for the $k$-NN classifier; several existing studies considered selecting optimal $k$ and weights to obtain faster convergence rate.

Stochastic Neighbor Embedding of Multimodal Relational Data for Image-Text Simultaneous Visualization

no code implementations2 May 2020 Morihiro Mizutani, Akifumi Okuno, Geewook Kim, Hidetoshi Shimodaira

Multimodal relational data analysis has become of increasing importance in recent years, for exploring across different domains of data, such as images and their text tags obtained from social networking services (e. g., Flickr).

Graph Embedding

Extrapolation Towards Imaginary $0$-Nearest Neighbour and Its Improved Convergence Rate

no code implementations8 Feb 2020 Akifumi Okuno, Hidetoshi Shimodaira

The weights and the parameter $k \in \mathbb{N}$ regulate its bias-variance trade-off, and the trade-off implicitly affects the convergence rate of the excess risk for the $k$-NN classifier; several existing studies considered selecting optimal $k$ and weights to obtain faster convergence rate.

More Powerful Selective Kernel Tests for Feature Selection

1 code implementation14 Oct 2019 Jen Ning Lim, Makoto Yamada, Wittawat Jitkrittum, Yoshikazu Terada, Shigeyuki Matsui, Hidetoshi Shimodaira

An approach for addressing this is via conditioning on the selection procedure to account for how we have used the data to generate our hypotheses, and prevent information to be used again after selection.

feature selection Selection bias

Hyperlink Regression via Bregman Divergence

no code implementations22 Jul 2019 Akifumi Okuno, Hidetoshi Shimodaira

A collection of $U \: (\in \mathbb{N})$ data vectors is called a $U$-tuple, and the association strength among the vectors of a tuple is termed as the \emph{hyperlink weight}, that is assumed to be symmetric with respect to permutation of the entries in the index.

Graph Embedding Link Prediction +4

Representation Learning with Weighted Inner Product for Universal Approximation of General Similarities

1 code implementation27 Feb 2019 Geewook Kim, Akifumi Okuno, Kazuki Fukui, Hidetoshi Shimodaira

In addition to the parameters of neural networks, we optimize the weights of the inner product by allowing positive and negative values.

Graph Embedding Model Selection +1

Robust Graph Embedding with Noisy Link Weights

no code implementations22 Feb 2019 Akifumi Okuno, Hidetoshi Shimodaira

We propose $\beta$-graph embedding for robustly learning feature vectors from data vectors and noisy link weights.

Graph Embedding

An information criterion for auxiliary variable selection in incomplete data analysis

no code implementations21 Feb 2019 Shinpei Imori, Hidetoshi Shimodaira

Utilizing a parametric model of joint distribution of primary and auxiliary variables, it is possible to improve the estimation of parametric model for the primary variables when the auxiliary variables are closely related to the primary variables.

Model Selection Variable Selection

Word-like character n-gram embedding

1 code implementation WS 2018 Geewook Kim, Kazuki Fukui, Hidetoshi Shimodaira

We propose a new word embedding method called \textit{word-like character} n\textit{-gram embedding}, which learns distributed representations of words by embedding word-like character n-grams.

Segmentation Word Embeddings

Graph Embedding with Shifted Inner Product Similarity and Its Improved Approximation Capability

no code implementations4 Oct 2018 Akifumi Okuno, Geewook Kim, Hidetoshi Shimodaira

We propose shifted inner-product similarity (SIPS), which is a novel yet very simple extension of the ordinary inner-product similarity (IPS) for neural-network based graph embedding (GE).

Graph Embedding

On representation power of neural network-based graph embedding and beyond

no code implementations31 May 2018 Akifumi Okuno, Hidetoshi Shimodaira

We consider the representation power of siamese-style similarity functions used in neural network-based graph embedding.

Graph Embedding

A probabilistic framework for multi-view feature learning with many-to-many associations via neural networks

no code implementations ICML 2018 Akifumi Okuno, Tetsuya Hada, Hidetoshi Shimodaira

PMvGE is a probabilistic model for predicting new associations via graph embedding of the nodes of data vectors with links of their associations.

Graph Embedding

Spectral Graph-Based Method of Multimodal Word Embedding

no code implementations WS 2017 Kazuki Fukui, Takamasa Oshikiri, Hidetoshi Shimodaira

In this paper, we propose a novel method for multimodal word embedding, which exploit a generalized framework of multi-view spectral graph embedding to take into account visual appearances or scenes denoted by words in a corpus.

Graph Embedding Image Retrieval +6

PAFit: an R Package for the Non-Parametric Estimation of Preferential Attachment and Node Fitness in Temporal Complex Networks

1 code implementation20 Apr 2017 Thong Pham, Paul Sheridan, Hidetoshi Shimodaira

This paper introduces the R package PAFit, which implements non-parametric procedures for estimating the preferential attachment function and node fitnesses in a growing network, as well as a number of functions for generating complex networks from these two mechanisms.

Data Analysis, Statistics and Probability Social and Information Networks Physics and Society Computation

Cross-validation of matching correlation analysis by resampling matching weights

no code implementations29 Mar 2015 Hidetoshi Shimodaira

For dimensionality reduction, we consider a linear transformation of data vectors, and define a matching error as the weighted sum of squared distances between transformed vectors with respect to the matching weights.

Dimensionality Reduction Graph Embedding

A simple coding for cross-domain matching with dimension reduction via spectral graph embedding

no code implementations29 Dec 2014 Hidetoshi Shimodaira

These data vectors from multiple domains are projected to a common space by linear transformations in order to search closely related vectors across domains.

Dimensionality Reduction Graph Embedding

Cannot find the paper you are looking for? You can Submit a new open access paper.