1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * Ceph cache definitions.
4*4882a593Smuzhiyun *
5*4882a593Smuzhiyun * Copyright (C) 2013 by Adfin Solutions, Inc. All Rights Reserved.
6*4882a593Smuzhiyun * Written by Milosz Tanski (milosz@adfin.com)
7*4882a593Smuzhiyun */
8*4882a593Smuzhiyun
9*4882a593Smuzhiyun #include <linux/ceph/ceph_debug.h>
10*4882a593Smuzhiyun
11*4882a593Smuzhiyun #include <linux/fs_context.h>
12*4882a593Smuzhiyun #include "super.h"
13*4882a593Smuzhiyun #include "cache.h"
14*4882a593Smuzhiyun
15*4882a593Smuzhiyun struct ceph_aux_inode {
16*4882a593Smuzhiyun u64 version;
17*4882a593Smuzhiyun u64 mtime_sec;
18*4882a593Smuzhiyun u64 mtime_nsec;
19*4882a593Smuzhiyun };
20*4882a593Smuzhiyun
21*4882a593Smuzhiyun struct fscache_netfs ceph_cache_netfs = {
22*4882a593Smuzhiyun .name = "ceph",
23*4882a593Smuzhiyun .version = 0,
24*4882a593Smuzhiyun };
25*4882a593Smuzhiyun
26*4882a593Smuzhiyun static DEFINE_MUTEX(ceph_fscache_lock);
27*4882a593Smuzhiyun static LIST_HEAD(ceph_fscache_list);
28*4882a593Smuzhiyun
29*4882a593Smuzhiyun struct ceph_fscache_entry {
30*4882a593Smuzhiyun struct list_head list;
31*4882a593Smuzhiyun struct fscache_cookie *fscache;
32*4882a593Smuzhiyun size_t uniq_len;
33*4882a593Smuzhiyun /* The following members must be last */
34*4882a593Smuzhiyun struct ceph_fsid fsid;
35*4882a593Smuzhiyun char uniquifier[];
36*4882a593Smuzhiyun };
37*4882a593Smuzhiyun
38*4882a593Smuzhiyun static const struct fscache_cookie_def ceph_fscache_fsid_object_def = {
39*4882a593Smuzhiyun .name = "CEPH.fsid",
40*4882a593Smuzhiyun .type = FSCACHE_COOKIE_TYPE_INDEX,
41*4882a593Smuzhiyun };
42*4882a593Smuzhiyun
ceph_fscache_register(void)43*4882a593Smuzhiyun int __init ceph_fscache_register(void)
44*4882a593Smuzhiyun {
45*4882a593Smuzhiyun return fscache_register_netfs(&ceph_cache_netfs);
46*4882a593Smuzhiyun }
47*4882a593Smuzhiyun
ceph_fscache_unregister(void)48*4882a593Smuzhiyun void ceph_fscache_unregister(void)
49*4882a593Smuzhiyun {
50*4882a593Smuzhiyun fscache_unregister_netfs(&ceph_cache_netfs);
51*4882a593Smuzhiyun }
52*4882a593Smuzhiyun
ceph_fscache_register_fs(struct ceph_fs_client * fsc,struct fs_context * fc)53*4882a593Smuzhiyun int ceph_fscache_register_fs(struct ceph_fs_client* fsc, struct fs_context *fc)
54*4882a593Smuzhiyun {
55*4882a593Smuzhiyun const struct ceph_fsid *fsid = &fsc->client->fsid;
56*4882a593Smuzhiyun const char *fscache_uniq = fsc->mount_options->fscache_uniq;
57*4882a593Smuzhiyun size_t uniq_len = fscache_uniq ? strlen(fscache_uniq) : 0;
58*4882a593Smuzhiyun struct ceph_fscache_entry *ent;
59*4882a593Smuzhiyun int err = 0;
60*4882a593Smuzhiyun
61*4882a593Smuzhiyun mutex_lock(&ceph_fscache_lock);
62*4882a593Smuzhiyun list_for_each_entry(ent, &ceph_fscache_list, list) {
63*4882a593Smuzhiyun if (memcmp(&ent->fsid, fsid, sizeof(*fsid)))
64*4882a593Smuzhiyun continue;
65*4882a593Smuzhiyun if (ent->uniq_len != uniq_len)
66*4882a593Smuzhiyun continue;
67*4882a593Smuzhiyun if (uniq_len && memcmp(ent->uniquifier, fscache_uniq, uniq_len))
68*4882a593Smuzhiyun continue;
69*4882a593Smuzhiyun
70*4882a593Smuzhiyun errorfc(fc, "fscache cookie already registered for fsid %pU, use fsc=<uniquifier> option",
71*4882a593Smuzhiyun fsid);
72*4882a593Smuzhiyun err = -EBUSY;
73*4882a593Smuzhiyun goto out_unlock;
74*4882a593Smuzhiyun }
75*4882a593Smuzhiyun
76*4882a593Smuzhiyun ent = kzalloc(sizeof(*ent) + uniq_len, GFP_KERNEL);
77*4882a593Smuzhiyun if (!ent) {
78*4882a593Smuzhiyun err = -ENOMEM;
79*4882a593Smuzhiyun goto out_unlock;
80*4882a593Smuzhiyun }
81*4882a593Smuzhiyun
82*4882a593Smuzhiyun memcpy(&ent->fsid, fsid, sizeof(*fsid));
83*4882a593Smuzhiyun if (uniq_len > 0) {
84*4882a593Smuzhiyun memcpy(&ent->uniquifier, fscache_uniq, uniq_len);
85*4882a593Smuzhiyun ent->uniq_len = uniq_len;
86*4882a593Smuzhiyun }
87*4882a593Smuzhiyun
88*4882a593Smuzhiyun fsc->fscache = fscache_acquire_cookie(ceph_cache_netfs.primary_index,
89*4882a593Smuzhiyun &ceph_fscache_fsid_object_def,
90*4882a593Smuzhiyun &ent->fsid, sizeof(ent->fsid) + uniq_len,
91*4882a593Smuzhiyun NULL, 0,
92*4882a593Smuzhiyun fsc, 0, true);
93*4882a593Smuzhiyun
94*4882a593Smuzhiyun if (fsc->fscache) {
95*4882a593Smuzhiyun ent->fscache = fsc->fscache;
96*4882a593Smuzhiyun list_add_tail(&ent->list, &ceph_fscache_list);
97*4882a593Smuzhiyun } else {
98*4882a593Smuzhiyun kfree(ent);
99*4882a593Smuzhiyun errorfc(fc, "unable to register fscache cookie for fsid %pU",
100*4882a593Smuzhiyun fsid);
101*4882a593Smuzhiyun /* all other fs ignore this error */
102*4882a593Smuzhiyun }
103*4882a593Smuzhiyun out_unlock:
104*4882a593Smuzhiyun mutex_unlock(&ceph_fscache_lock);
105*4882a593Smuzhiyun return err;
106*4882a593Smuzhiyun }
107*4882a593Smuzhiyun
ceph_fscache_inode_check_aux(void * cookie_netfs_data,const void * data,uint16_t dlen,loff_t object_size)108*4882a593Smuzhiyun static enum fscache_checkaux ceph_fscache_inode_check_aux(
109*4882a593Smuzhiyun void *cookie_netfs_data, const void *data, uint16_t dlen,
110*4882a593Smuzhiyun loff_t object_size)
111*4882a593Smuzhiyun {
112*4882a593Smuzhiyun struct ceph_aux_inode aux;
113*4882a593Smuzhiyun struct ceph_inode_info* ci = cookie_netfs_data;
114*4882a593Smuzhiyun struct inode* inode = &ci->vfs_inode;
115*4882a593Smuzhiyun
116*4882a593Smuzhiyun if (dlen != sizeof(aux) ||
117*4882a593Smuzhiyun i_size_read(inode) != object_size)
118*4882a593Smuzhiyun return FSCACHE_CHECKAUX_OBSOLETE;
119*4882a593Smuzhiyun
120*4882a593Smuzhiyun memset(&aux, 0, sizeof(aux));
121*4882a593Smuzhiyun aux.version = ci->i_version;
122*4882a593Smuzhiyun aux.mtime_sec = inode->i_mtime.tv_sec;
123*4882a593Smuzhiyun aux.mtime_nsec = inode->i_mtime.tv_nsec;
124*4882a593Smuzhiyun
125*4882a593Smuzhiyun if (memcmp(data, &aux, sizeof(aux)) != 0)
126*4882a593Smuzhiyun return FSCACHE_CHECKAUX_OBSOLETE;
127*4882a593Smuzhiyun
128*4882a593Smuzhiyun dout("ceph inode 0x%p cached okay\n", ci);
129*4882a593Smuzhiyun return FSCACHE_CHECKAUX_OKAY;
130*4882a593Smuzhiyun }
131*4882a593Smuzhiyun
132*4882a593Smuzhiyun static const struct fscache_cookie_def ceph_fscache_inode_object_def = {
133*4882a593Smuzhiyun .name = "CEPH.inode",
134*4882a593Smuzhiyun .type = FSCACHE_COOKIE_TYPE_DATAFILE,
135*4882a593Smuzhiyun .check_aux = ceph_fscache_inode_check_aux,
136*4882a593Smuzhiyun };
137*4882a593Smuzhiyun
ceph_fscache_register_inode_cookie(struct inode * inode)138*4882a593Smuzhiyun void ceph_fscache_register_inode_cookie(struct inode *inode)
139*4882a593Smuzhiyun {
140*4882a593Smuzhiyun struct ceph_inode_info *ci = ceph_inode(inode);
141*4882a593Smuzhiyun struct ceph_fs_client *fsc = ceph_inode_to_client(inode);
142*4882a593Smuzhiyun struct ceph_aux_inode aux;
143*4882a593Smuzhiyun
144*4882a593Smuzhiyun /* No caching for filesystem */
145*4882a593Smuzhiyun if (!fsc->fscache)
146*4882a593Smuzhiyun return;
147*4882a593Smuzhiyun
148*4882a593Smuzhiyun /* Only cache for regular files that are read only */
149*4882a593Smuzhiyun if (!S_ISREG(inode->i_mode))
150*4882a593Smuzhiyun return;
151*4882a593Smuzhiyun
152*4882a593Smuzhiyun inode_lock_nested(inode, I_MUTEX_CHILD);
153*4882a593Smuzhiyun if (!ci->fscache) {
154*4882a593Smuzhiyun memset(&aux, 0, sizeof(aux));
155*4882a593Smuzhiyun aux.version = ci->i_version;
156*4882a593Smuzhiyun aux.mtime_sec = inode->i_mtime.tv_sec;
157*4882a593Smuzhiyun aux.mtime_nsec = inode->i_mtime.tv_nsec;
158*4882a593Smuzhiyun ci->fscache = fscache_acquire_cookie(fsc->fscache,
159*4882a593Smuzhiyun &ceph_fscache_inode_object_def,
160*4882a593Smuzhiyun &ci->i_vino, sizeof(ci->i_vino),
161*4882a593Smuzhiyun &aux, sizeof(aux),
162*4882a593Smuzhiyun ci, i_size_read(inode), false);
163*4882a593Smuzhiyun }
164*4882a593Smuzhiyun inode_unlock(inode);
165*4882a593Smuzhiyun }
166*4882a593Smuzhiyun
ceph_fscache_unregister_inode_cookie(struct ceph_inode_info * ci)167*4882a593Smuzhiyun void ceph_fscache_unregister_inode_cookie(struct ceph_inode_info* ci)
168*4882a593Smuzhiyun {
169*4882a593Smuzhiyun struct fscache_cookie* cookie;
170*4882a593Smuzhiyun
171*4882a593Smuzhiyun if ((cookie = ci->fscache) == NULL)
172*4882a593Smuzhiyun return;
173*4882a593Smuzhiyun
174*4882a593Smuzhiyun ci->fscache = NULL;
175*4882a593Smuzhiyun
176*4882a593Smuzhiyun fscache_uncache_all_inode_pages(cookie, &ci->vfs_inode);
177*4882a593Smuzhiyun fscache_relinquish_cookie(cookie, &ci->i_vino, false);
178*4882a593Smuzhiyun }
179*4882a593Smuzhiyun
ceph_fscache_can_enable(void * data)180*4882a593Smuzhiyun static bool ceph_fscache_can_enable(void *data)
181*4882a593Smuzhiyun {
182*4882a593Smuzhiyun struct inode *inode = data;
183*4882a593Smuzhiyun return !inode_is_open_for_write(inode);
184*4882a593Smuzhiyun }
185*4882a593Smuzhiyun
ceph_fscache_file_set_cookie(struct inode * inode,struct file * filp)186*4882a593Smuzhiyun void ceph_fscache_file_set_cookie(struct inode *inode, struct file *filp)
187*4882a593Smuzhiyun {
188*4882a593Smuzhiyun struct ceph_inode_info *ci = ceph_inode(inode);
189*4882a593Smuzhiyun
190*4882a593Smuzhiyun if (!fscache_cookie_valid(ci->fscache))
191*4882a593Smuzhiyun return;
192*4882a593Smuzhiyun
193*4882a593Smuzhiyun if (inode_is_open_for_write(inode)) {
194*4882a593Smuzhiyun dout("fscache_file_set_cookie %p %p disabling cache\n",
195*4882a593Smuzhiyun inode, filp);
196*4882a593Smuzhiyun fscache_disable_cookie(ci->fscache, &ci->i_vino, false);
197*4882a593Smuzhiyun fscache_uncache_all_inode_pages(ci->fscache, inode);
198*4882a593Smuzhiyun } else {
199*4882a593Smuzhiyun fscache_enable_cookie(ci->fscache, &ci->i_vino, i_size_read(inode),
200*4882a593Smuzhiyun ceph_fscache_can_enable, inode);
201*4882a593Smuzhiyun if (fscache_cookie_enabled(ci->fscache)) {
202*4882a593Smuzhiyun dout("fscache_file_set_cookie %p %p enabling cache\n",
203*4882a593Smuzhiyun inode, filp);
204*4882a593Smuzhiyun }
205*4882a593Smuzhiyun }
206*4882a593Smuzhiyun }
207*4882a593Smuzhiyun
ceph_readpage_from_fscache_complete(struct page * page,void * data,int error)208*4882a593Smuzhiyun static void ceph_readpage_from_fscache_complete(struct page *page, void *data, int error)
209*4882a593Smuzhiyun {
210*4882a593Smuzhiyun if (!error)
211*4882a593Smuzhiyun SetPageUptodate(page);
212*4882a593Smuzhiyun
213*4882a593Smuzhiyun unlock_page(page);
214*4882a593Smuzhiyun }
215*4882a593Smuzhiyun
cache_valid(struct ceph_inode_info * ci)216*4882a593Smuzhiyun static inline bool cache_valid(struct ceph_inode_info *ci)
217*4882a593Smuzhiyun {
218*4882a593Smuzhiyun return ci->i_fscache_gen == ci->i_rdcache_gen;
219*4882a593Smuzhiyun }
220*4882a593Smuzhiyun
221*4882a593Smuzhiyun
222*4882a593Smuzhiyun /* Atempt to read from the fscache,
223*4882a593Smuzhiyun *
224*4882a593Smuzhiyun * This function is called from the readpage_nounlock context. DO NOT attempt to
225*4882a593Smuzhiyun * unlock the page here (or in the callback).
226*4882a593Smuzhiyun */
ceph_readpage_from_fscache(struct inode * inode,struct page * page)227*4882a593Smuzhiyun int ceph_readpage_from_fscache(struct inode *inode, struct page *page)
228*4882a593Smuzhiyun {
229*4882a593Smuzhiyun struct ceph_inode_info *ci = ceph_inode(inode);
230*4882a593Smuzhiyun int ret;
231*4882a593Smuzhiyun
232*4882a593Smuzhiyun if (!cache_valid(ci))
233*4882a593Smuzhiyun return -ENOBUFS;
234*4882a593Smuzhiyun
235*4882a593Smuzhiyun ret = fscache_read_or_alloc_page(ci->fscache, page,
236*4882a593Smuzhiyun ceph_readpage_from_fscache_complete, NULL,
237*4882a593Smuzhiyun GFP_KERNEL);
238*4882a593Smuzhiyun
239*4882a593Smuzhiyun switch (ret) {
240*4882a593Smuzhiyun case 0: /* Page found */
241*4882a593Smuzhiyun dout("page read submitted\n");
242*4882a593Smuzhiyun return 0;
243*4882a593Smuzhiyun case -ENOBUFS: /* Pages were not found, and can't be */
244*4882a593Smuzhiyun case -ENODATA: /* Pages were not found */
245*4882a593Smuzhiyun dout("page/inode not in cache\n");
246*4882a593Smuzhiyun return ret;
247*4882a593Smuzhiyun default:
248*4882a593Smuzhiyun dout("%s: unknown error ret = %i\n", __func__, ret);
249*4882a593Smuzhiyun return ret;
250*4882a593Smuzhiyun }
251*4882a593Smuzhiyun }
252*4882a593Smuzhiyun
ceph_readpages_from_fscache(struct inode * inode,struct address_space * mapping,struct list_head * pages,unsigned * nr_pages)253*4882a593Smuzhiyun int ceph_readpages_from_fscache(struct inode *inode,
254*4882a593Smuzhiyun struct address_space *mapping,
255*4882a593Smuzhiyun struct list_head *pages,
256*4882a593Smuzhiyun unsigned *nr_pages)
257*4882a593Smuzhiyun {
258*4882a593Smuzhiyun struct ceph_inode_info *ci = ceph_inode(inode);
259*4882a593Smuzhiyun int ret;
260*4882a593Smuzhiyun
261*4882a593Smuzhiyun if (!cache_valid(ci))
262*4882a593Smuzhiyun return -ENOBUFS;
263*4882a593Smuzhiyun
264*4882a593Smuzhiyun ret = fscache_read_or_alloc_pages(ci->fscache, mapping, pages, nr_pages,
265*4882a593Smuzhiyun ceph_readpage_from_fscache_complete,
266*4882a593Smuzhiyun NULL, mapping_gfp_mask(mapping));
267*4882a593Smuzhiyun
268*4882a593Smuzhiyun switch (ret) {
269*4882a593Smuzhiyun case 0: /* All pages found */
270*4882a593Smuzhiyun dout("all-page read submitted\n");
271*4882a593Smuzhiyun return 0;
272*4882a593Smuzhiyun case -ENOBUFS: /* Some pages were not found, and can't be */
273*4882a593Smuzhiyun case -ENODATA: /* some pages were not found */
274*4882a593Smuzhiyun dout("page/inode not in cache\n");
275*4882a593Smuzhiyun return ret;
276*4882a593Smuzhiyun default:
277*4882a593Smuzhiyun dout("%s: unknown error ret = %i\n", __func__, ret);
278*4882a593Smuzhiyun return ret;
279*4882a593Smuzhiyun }
280*4882a593Smuzhiyun }
281*4882a593Smuzhiyun
ceph_readpage_to_fscache(struct inode * inode,struct page * page)282*4882a593Smuzhiyun void ceph_readpage_to_fscache(struct inode *inode, struct page *page)
283*4882a593Smuzhiyun {
284*4882a593Smuzhiyun struct ceph_inode_info *ci = ceph_inode(inode);
285*4882a593Smuzhiyun int ret;
286*4882a593Smuzhiyun
287*4882a593Smuzhiyun if (!PageFsCache(page))
288*4882a593Smuzhiyun return;
289*4882a593Smuzhiyun
290*4882a593Smuzhiyun if (!cache_valid(ci))
291*4882a593Smuzhiyun return;
292*4882a593Smuzhiyun
293*4882a593Smuzhiyun ret = fscache_write_page(ci->fscache, page, i_size_read(inode),
294*4882a593Smuzhiyun GFP_KERNEL);
295*4882a593Smuzhiyun if (ret)
296*4882a593Smuzhiyun fscache_uncache_page(ci->fscache, page);
297*4882a593Smuzhiyun }
298*4882a593Smuzhiyun
ceph_invalidate_fscache_page(struct inode * inode,struct page * page)299*4882a593Smuzhiyun void ceph_invalidate_fscache_page(struct inode* inode, struct page *page)
300*4882a593Smuzhiyun {
301*4882a593Smuzhiyun struct ceph_inode_info *ci = ceph_inode(inode);
302*4882a593Smuzhiyun
303*4882a593Smuzhiyun if (!PageFsCache(page))
304*4882a593Smuzhiyun return;
305*4882a593Smuzhiyun
306*4882a593Smuzhiyun fscache_wait_on_page_write(ci->fscache, page);
307*4882a593Smuzhiyun fscache_uncache_page(ci->fscache, page);
308*4882a593Smuzhiyun }
309*4882a593Smuzhiyun
ceph_fscache_unregister_fs(struct ceph_fs_client * fsc)310*4882a593Smuzhiyun void ceph_fscache_unregister_fs(struct ceph_fs_client* fsc)
311*4882a593Smuzhiyun {
312*4882a593Smuzhiyun if (fscache_cookie_valid(fsc->fscache)) {
313*4882a593Smuzhiyun struct ceph_fscache_entry *ent;
314*4882a593Smuzhiyun bool found = false;
315*4882a593Smuzhiyun
316*4882a593Smuzhiyun mutex_lock(&ceph_fscache_lock);
317*4882a593Smuzhiyun list_for_each_entry(ent, &ceph_fscache_list, list) {
318*4882a593Smuzhiyun if (ent->fscache == fsc->fscache) {
319*4882a593Smuzhiyun list_del(&ent->list);
320*4882a593Smuzhiyun kfree(ent);
321*4882a593Smuzhiyun found = true;
322*4882a593Smuzhiyun break;
323*4882a593Smuzhiyun }
324*4882a593Smuzhiyun }
325*4882a593Smuzhiyun WARN_ON_ONCE(!found);
326*4882a593Smuzhiyun mutex_unlock(&ceph_fscache_lock);
327*4882a593Smuzhiyun
328*4882a593Smuzhiyun __fscache_relinquish_cookie(fsc->fscache, NULL, false);
329*4882a593Smuzhiyun }
330*4882a593Smuzhiyun fsc->fscache = NULL;
331*4882a593Smuzhiyun }
332*4882a593Smuzhiyun
333*4882a593Smuzhiyun /*
334*4882a593Smuzhiyun * caller should hold CEPH_CAP_FILE_{RD,CACHE}
335*4882a593Smuzhiyun */
ceph_fscache_revalidate_cookie(struct ceph_inode_info * ci)336*4882a593Smuzhiyun void ceph_fscache_revalidate_cookie(struct ceph_inode_info *ci)
337*4882a593Smuzhiyun {
338*4882a593Smuzhiyun if (cache_valid(ci))
339*4882a593Smuzhiyun return;
340*4882a593Smuzhiyun
341*4882a593Smuzhiyun /* resue i_truncate_mutex. There should be no pending
342*4882a593Smuzhiyun * truncate while the caller holds CEPH_CAP_FILE_RD */
343*4882a593Smuzhiyun mutex_lock(&ci->i_truncate_mutex);
344*4882a593Smuzhiyun if (!cache_valid(ci)) {
345*4882a593Smuzhiyun if (fscache_check_consistency(ci->fscache, &ci->i_vino))
346*4882a593Smuzhiyun fscache_invalidate(ci->fscache);
347*4882a593Smuzhiyun spin_lock(&ci->i_ceph_lock);
348*4882a593Smuzhiyun ci->i_fscache_gen = ci->i_rdcache_gen;
349*4882a593Smuzhiyun spin_unlock(&ci->i_ceph_lock);
350*4882a593Smuzhiyun }
351*4882a593Smuzhiyun mutex_unlock(&ci->i_truncate_mutex);
352*4882a593Smuzhiyun }
353