|  | /* | 
|  | * Copyright (c) 2000-2005 Silicon Graphics, Inc. | 
|  | * All Rights Reserved. | 
|  | * | 
|  | * This program is free software; you can redistribute it and/or | 
|  | * modify it under the terms of the GNU General Public License as | 
|  | * published by the Free Software Foundation. | 
|  | * | 
|  | * This program is distributed in the hope that it would be useful, | 
|  | * but WITHOUT ANY WARRANTY; without even the implied warranty of | 
|  | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the | 
|  | * GNU General Public License for more details. | 
|  | * | 
|  | * You should have received a copy of the GNU General Public License | 
|  | * along with this program; if not, write the Free Software Foundation, | 
|  | * Inc.,  51 Franklin St, Fifth Floor, Boston, MA  02110-1301  USA | 
|  | */ | 
|  |  | 
|  | #include <linux/capability.h> | 
|  |  | 
|  | #include "xfs.h" | 
|  | #include "xfs_fs.h" | 
|  | #include "xfs_shared.h" | 
|  | #include "xfs_format.h" | 
|  | #include "xfs_log_format.h" | 
|  | #include "xfs_trans_resv.h" | 
|  | #include "xfs_bit.h" | 
|  | #include "xfs_sb.h" | 
|  | #include "xfs_mount.h" | 
|  | #include "xfs_inode.h" | 
|  | #include "xfs_trans.h" | 
|  | #include "xfs_error.h" | 
|  | #include "xfs_quota.h" | 
|  | #include "xfs_qm.h" | 
|  | #include "xfs_trace.h" | 
|  | #include "xfs_icache.h" | 
|  |  | 
|  | STATIC int	xfs_qm_log_quotaoff(xfs_mount_t *, xfs_qoff_logitem_t **, uint); | 
|  | STATIC int	xfs_qm_log_quotaoff_end(xfs_mount_t *, xfs_qoff_logitem_t *, | 
|  | uint); | 
|  | STATIC uint	xfs_qm_export_flags(uint); | 
|  |  | 
|  | /* | 
|  | * Turn off quota accounting and/or enforcement for all udquots and/or | 
|  | * gdquots. Called only at unmount time. | 
|  | * | 
|  | * This assumes that there are no dquots of this file system cached | 
|  | * incore, and modifies the ondisk dquot directly. Therefore, for example, | 
|  | * it is an error to call this twice, without purging the cache. | 
|  | */ | 
|  | int | 
|  | xfs_qm_scall_quotaoff( | 
|  | xfs_mount_t		*mp, | 
|  | uint			flags) | 
|  | { | 
|  | struct xfs_quotainfo	*q = mp->m_quotainfo; | 
|  | uint			dqtype; | 
|  | int			error; | 
|  | uint			inactivate_flags; | 
|  | xfs_qoff_logitem_t	*qoffstart; | 
|  |  | 
|  | /* | 
|  | * No file system can have quotas enabled on disk but not in core. | 
|  | * Note that quota utilities (like quotaoff) _expect_ | 
|  | * errno == -EEXIST here. | 
|  | */ | 
|  | if ((mp->m_qflags & flags) == 0) | 
|  | return -EEXIST; | 
|  | error = 0; | 
|  |  | 
|  | flags &= (XFS_ALL_QUOTA_ACCT | XFS_ALL_QUOTA_ENFD); | 
|  |  | 
|  | /* | 
|  | * We don't want to deal with two quotaoffs messing up each other, | 
|  | * so we're going to serialize it. quotaoff isn't exactly a performance | 
|  | * critical thing. | 
|  | * If quotaoff, then we must be dealing with the root filesystem. | 
|  | */ | 
|  | ASSERT(q); | 
|  | mutex_lock(&q->qi_quotaofflock); | 
|  |  | 
|  | /* | 
|  | * If we're just turning off quota enforcement, change mp and go. | 
|  | */ | 
|  | if ((flags & XFS_ALL_QUOTA_ACCT) == 0) { | 
|  | mp->m_qflags &= ~(flags); | 
|  |  | 
|  | spin_lock(&mp->m_sb_lock); | 
|  | mp->m_sb.sb_qflags = mp->m_qflags; | 
|  | spin_unlock(&mp->m_sb_lock); | 
|  | mutex_unlock(&q->qi_quotaofflock); | 
|  |  | 
|  | /* XXX what to do if error ? Revert back to old vals incore ? */ | 
|  | return xfs_sync_sb(mp, false); | 
|  | } | 
|  |  | 
|  | dqtype = 0; | 
|  | inactivate_flags = 0; | 
|  | /* | 
|  | * If accounting is off, we must turn enforcement off, clear the | 
|  | * quota 'CHKD' certificate to make it known that we have to | 
|  | * do a quotacheck the next time this quota is turned on. | 
|  | */ | 
|  | if (flags & XFS_UQUOTA_ACCT) { | 
|  | dqtype |= XFS_QMOPT_UQUOTA; | 
|  | flags |= (XFS_UQUOTA_CHKD | XFS_UQUOTA_ENFD); | 
|  | inactivate_flags |= XFS_UQUOTA_ACTIVE; | 
|  | } | 
|  | if (flags & XFS_GQUOTA_ACCT) { | 
|  | dqtype |= XFS_QMOPT_GQUOTA; | 
|  | flags |= (XFS_GQUOTA_CHKD | XFS_GQUOTA_ENFD); | 
|  | inactivate_flags |= XFS_GQUOTA_ACTIVE; | 
|  | } | 
|  | if (flags & XFS_PQUOTA_ACCT) { | 
|  | dqtype |= XFS_QMOPT_PQUOTA; | 
|  | flags |= (XFS_PQUOTA_CHKD | XFS_PQUOTA_ENFD); | 
|  | inactivate_flags |= XFS_PQUOTA_ACTIVE; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Nothing to do?  Don't complain. This happens when we're just | 
|  | * turning off quota enforcement. | 
|  | */ | 
|  | if ((mp->m_qflags & flags) == 0) | 
|  | goto out_unlock; | 
|  |  | 
|  | /* | 
|  | * Write the LI_QUOTAOFF log record, and do SB changes atomically, | 
|  | * and synchronously. If we fail to write, we should abort the | 
|  | * operation as it cannot be recovered safely if we crash. | 
|  | */ | 
|  | error = xfs_qm_log_quotaoff(mp, &qoffstart, flags); | 
|  | if (error) | 
|  | goto out_unlock; | 
|  |  | 
|  | /* | 
|  | * Next we clear the XFS_MOUNT_*DQ_ACTIVE bit(s) in the mount struct | 
|  | * to take care of the race between dqget and quotaoff. We don't take | 
|  | * any special locks to reset these bits. All processes need to check | 
|  | * these bits *after* taking inode lock(s) to see if the particular | 
|  | * quota type is in the process of being turned off. If *ACTIVE, it is | 
|  | * guaranteed that all dquot structures and all quotainode ptrs will all | 
|  | * stay valid as long as that inode is kept locked. | 
|  | * | 
|  | * There is no turning back after this. | 
|  | */ | 
|  | mp->m_qflags &= ~inactivate_flags; | 
|  |  | 
|  | /* | 
|  | * Give back all the dquot reference(s) held by inodes. | 
|  | * Here we go thru every single incore inode in this file system, and | 
|  | * do a dqrele on the i_udquot/i_gdquot that it may have. | 
|  | * Essentially, as long as somebody has an inode locked, this guarantees | 
|  | * that quotas will not be turned off. This is handy because in a | 
|  | * transaction once we lock the inode(s) and check for quotaon, we can | 
|  | * depend on the quota inodes (and other things) being valid as long as | 
|  | * we keep the lock(s). | 
|  | */ | 
|  | xfs_qm_dqrele_all_inodes(mp, flags); | 
|  |  | 
|  | /* | 
|  | * Next we make the changes in the quota flag in the mount struct. | 
|  | * This isn't protected by a particular lock directly, because we | 
|  | * don't want to take a mrlock every time we depend on quotas being on. | 
|  | */ | 
|  | mp->m_qflags &= ~flags; | 
|  |  | 
|  | /* | 
|  | * Go through all the dquots of this file system and purge them, | 
|  | * according to what was turned off. | 
|  | */ | 
|  | xfs_qm_dqpurge_all(mp, dqtype); | 
|  |  | 
|  | /* | 
|  | * Transactions that had started before ACTIVE state bit was cleared | 
|  | * could have logged many dquots, so they'd have higher LSNs than | 
|  | * the first QUOTAOFF log record does. If we happen to crash when | 
|  | * the tail of the log has gone past the QUOTAOFF record, but | 
|  | * before the last dquot modification, those dquots __will__ | 
|  | * recover, and that's not good. | 
|  | * | 
|  | * So, we have QUOTAOFF start and end logitems; the start | 
|  | * logitem won't get overwritten until the end logitem appears... | 
|  | */ | 
|  | error = xfs_qm_log_quotaoff_end(mp, qoffstart, flags); | 
|  | if (error) { | 
|  | /* We're screwed now. Shutdown is the only option. */ | 
|  | xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE); | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * If all quotas are completely turned off, close shop. | 
|  | */ | 
|  | if (mp->m_qflags == 0) { | 
|  | mutex_unlock(&q->qi_quotaofflock); | 
|  | xfs_qm_destroy_quotainfo(mp); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Release our quotainode references if we don't need them anymore. | 
|  | */ | 
|  | if ((dqtype & XFS_QMOPT_UQUOTA) && q->qi_uquotaip) { | 
|  | IRELE(q->qi_uquotaip); | 
|  | q->qi_uquotaip = NULL; | 
|  | } | 
|  | if ((dqtype & XFS_QMOPT_GQUOTA) && q->qi_gquotaip) { | 
|  | IRELE(q->qi_gquotaip); | 
|  | q->qi_gquotaip = NULL; | 
|  | } | 
|  | if ((dqtype & XFS_QMOPT_PQUOTA) && q->qi_pquotaip) { | 
|  | IRELE(q->qi_pquotaip); | 
|  | q->qi_pquotaip = NULL; | 
|  | } | 
|  |  | 
|  | out_unlock: | 
|  | mutex_unlock(&q->qi_quotaofflock); | 
|  | return error; | 
|  | } | 
|  |  | 
|  | STATIC int | 
|  | xfs_qm_scall_trunc_qfile( | 
|  | struct xfs_mount	*mp, | 
|  | xfs_ino_t		ino) | 
|  | { | 
|  | struct xfs_inode	*ip; | 
|  | struct xfs_trans	*tp; | 
|  | int			error; | 
|  |  | 
|  | if (ino == NULLFSINO) | 
|  | return 0; | 
|  |  | 
|  | error = xfs_iget(mp, NULL, ino, 0, 0, &ip); | 
|  | if (error) | 
|  | return error; | 
|  |  | 
|  | xfs_ilock(ip, XFS_IOLOCK_EXCL); | 
|  |  | 
|  | tp = xfs_trans_alloc(mp, XFS_TRANS_TRUNCATE_FILE); | 
|  | error = xfs_trans_reserve(tp, &M_RES(mp)->tr_itruncate, 0, 0); | 
|  | if (error) { | 
|  | xfs_trans_cancel(tp, 0); | 
|  | xfs_iunlock(ip, XFS_IOLOCK_EXCL); | 
|  | goto out_put; | 
|  | } | 
|  |  | 
|  | xfs_ilock(ip, XFS_ILOCK_EXCL); | 
|  | xfs_trans_ijoin(tp, ip, 0); | 
|  |  | 
|  | ip->i_d.di_size = 0; | 
|  | xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE); | 
|  |  | 
|  | error = xfs_itruncate_extents(&tp, ip, XFS_DATA_FORK, 0); | 
|  | if (error) { | 
|  | xfs_trans_cancel(tp, XFS_TRANS_RELEASE_LOG_RES | | 
|  | XFS_TRANS_ABORT); | 
|  | goto out_unlock; | 
|  | } | 
|  |  | 
|  | ASSERT(ip->i_d.di_nextents == 0); | 
|  |  | 
|  | xfs_trans_ichgtime(tp, ip, XFS_ICHGTIME_MOD | XFS_ICHGTIME_CHG); | 
|  | error = xfs_trans_commit(tp, XFS_TRANS_RELEASE_LOG_RES); | 
|  |  | 
|  | out_unlock: | 
|  | xfs_iunlock(ip, XFS_ILOCK_EXCL | XFS_IOLOCK_EXCL); | 
|  | out_put: | 
|  | IRELE(ip); | 
|  | return error; | 
|  | } | 
|  |  | 
|  | int | 
|  | xfs_qm_scall_trunc_qfiles( | 
|  | xfs_mount_t	*mp, | 
|  | uint		flags) | 
|  | { | 
|  | int		error = -EINVAL; | 
|  |  | 
|  | if (!xfs_sb_version_hasquota(&mp->m_sb) || flags == 0 || | 
|  | (flags & ~XFS_DQ_ALLTYPES)) { | 
|  | xfs_debug(mp, "%s: flags=%x m_qflags=%x", | 
|  | __func__, flags, mp->m_qflags); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | if (flags & XFS_DQ_USER) { | 
|  | error = xfs_qm_scall_trunc_qfile(mp, mp->m_sb.sb_uquotino); | 
|  | if (error) | 
|  | return error; | 
|  | } | 
|  | if (flags & XFS_DQ_GROUP) { | 
|  | error = xfs_qm_scall_trunc_qfile(mp, mp->m_sb.sb_gquotino); | 
|  | if (error) | 
|  | return error; | 
|  | } | 
|  | if (flags & XFS_DQ_PROJ) | 
|  | error = xfs_qm_scall_trunc_qfile(mp, mp->m_sb.sb_pquotino); | 
|  |  | 
|  | return error; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Switch on (a given) quota enforcement for a filesystem.  This takes | 
|  | * effect immediately. | 
|  | * (Switching on quota accounting must be done at mount time.) | 
|  | */ | 
|  | int | 
|  | xfs_qm_scall_quotaon( | 
|  | xfs_mount_t	*mp, | 
|  | uint		flags) | 
|  | { | 
|  | int		error; | 
|  | uint		qf; | 
|  |  | 
|  | flags &= (XFS_ALL_QUOTA_ACCT | XFS_ALL_QUOTA_ENFD); | 
|  | /* | 
|  | * Switching on quota accounting must be done at mount time. | 
|  | */ | 
|  | flags &= ~(XFS_ALL_QUOTA_ACCT); | 
|  |  | 
|  | if (flags == 0) { | 
|  | xfs_debug(mp, "%s: zero flags, m_qflags=%x", | 
|  | __func__, mp->m_qflags); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Can't enforce without accounting. We check the superblock | 
|  | * qflags here instead of m_qflags because rootfs can have | 
|  | * quota acct on ondisk without m_qflags' knowing. | 
|  | */ | 
|  | if (((mp->m_sb.sb_qflags & XFS_UQUOTA_ACCT) == 0 && | 
|  | (flags & XFS_UQUOTA_ENFD)) || | 
|  | ((mp->m_sb.sb_qflags & XFS_GQUOTA_ACCT) == 0 && | 
|  | (flags & XFS_GQUOTA_ENFD)) || | 
|  | ((mp->m_sb.sb_qflags & XFS_PQUOTA_ACCT) == 0 && | 
|  | (flags & XFS_PQUOTA_ENFD))) { | 
|  | xfs_debug(mp, | 
|  | "%s: Can't enforce without acct, flags=%x sbflags=%x", | 
|  | __func__, flags, mp->m_sb.sb_qflags); | 
|  | return -EINVAL; | 
|  | } | 
|  | /* | 
|  | * If everything's up to-date incore, then don't waste time. | 
|  | */ | 
|  | if ((mp->m_qflags & flags) == flags) | 
|  | return -EEXIST; | 
|  |  | 
|  | /* | 
|  | * Change sb_qflags on disk but not incore mp->qflags | 
|  | * if this is the root filesystem. | 
|  | */ | 
|  | spin_lock(&mp->m_sb_lock); | 
|  | qf = mp->m_sb.sb_qflags; | 
|  | mp->m_sb.sb_qflags = qf | flags; | 
|  | spin_unlock(&mp->m_sb_lock); | 
|  |  | 
|  | /* | 
|  | * There's nothing to change if it's the same. | 
|  | */ | 
|  | if ((qf & flags) == flags) | 
|  | return -EEXIST; | 
|  |  | 
|  | error = xfs_sync_sb(mp, false); | 
|  | if (error) | 
|  | return error; | 
|  | /* | 
|  | * If we aren't trying to switch on quota enforcement, we are done. | 
|  | */ | 
|  | if  (((mp->m_sb.sb_qflags & XFS_UQUOTA_ACCT) != | 
|  | (mp->m_qflags & XFS_UQUOTA_ACCT)) || | 
|  | ((mp->m_sb.sb_qflags & XFS_PQUOTA_ACCT) != | 
|  | (mp->m_qflags & XFS_PQUOTA_ACCT)) || | 
|  | ((mp->m_sb.sb_qflags & XFS_GQUOTA_ACCT) != | 
|  | (mp->m_qflags & XFS_GQUOTA_ACCT))) | 
|  | return 0; | 
|  |  | 
|  | if (! XFS_IS_QUOTA_RUNNING(mp)) | 
|  | return -ESRCH; | 
|  |  | 
|  | /* | 
|  | * Switch on quota enforcement in core. | 
|  | */ | 
|  | mutex_lock(&mp->m_quotainfo->qi_quotaofflock); | 
|  | mp->m_qflags |= (flags & XFS_ALL_QUOTA_ENFD); | 
|  | mutex_unlock(&mp->m_quotainfo->qi_quotaofflock); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  |  | 
|  | /* | 
|  | * Return quota status information, such as uquota-off, enforcements, etc. | 
|  | * for Q_XGETQSTAT command. | 
|  | */ | 
|  | int | 
|  | xfs_qm_scall_getqstat( | 
|  | struct xfs_mount	*mp, | 
|  | struct fs_quota_stat	*out) | 
|  | { | 
|  | struct xfs_quotainfo	*q = mp->m_quotainfo; | 
|  | struct xfs_inode	*uip = NULL; | 
|  | struct xfs_inode	*gip = NULL; | 
|  | struct xfs_inode	*pip = NULL; | 
|  | bool                    tempuqip = false; | 
|  | bool                    tempgqip = false; | 
|  | bool                    temppqip = false; | 
|  |  | 
|  | memset(out, 0, sizeof(fs_quota_stat_t)); | 
|  |  | 
|  | out->qs_version = FS_QSTAT_VERSION; | 
|  | out->qs_flags = (__uint16_t) xfs_qm_export_flags(mp->m_qflags & | 
|  | (XFS_ALL_QUOTA_ACCT| | 
|  | XFS_ALL_QUOTA_ENFD)); | 
|  | uip = q->qi_uquotaip; | 
|  | gip = q->qi_gquotaip; | 
|  | pip = q->qi_pquotaip; | 
|  | if (!uip && mp->m_sb.sb_uquotino != NULLFSINO) { | 
|  | if (xfs_iget(mp, NULL, mp->m_sb.sb_uquotino, | 
|  | 0, 0, &uip) == 0) | 
|  | tempuqip = true; | 
|  | } | 
|  | if (!gip && mp->m_sb.sb_gquotino != NULLFSINO) { | 
|  | if (xfs_iget(mp, NULL, mp->m_sb.sb_gquotino, | 
|  | 0, 0, &gip) == 0) | 
|  | tempgqip = true; | 
|  | } | 
|  | /* | 
|  | * Q_XGETQSTAT doesn't have room for both group and project quotas. | 
|  | * So, allow the project quota values to be copied out only if | 
|  | * there is no group quota information available. | 
|  | */ | 
|  | if (!gip) { | 
|  | if (!pip && mp->m_sb.sb_pquotino != NULLFSINO) { | 
|  | if (xfs_iget(mp, NULL, mp->m_sb.sb_pquotino, | 
|  | 0, 0, &pip) == 0) | 
|  | temppqip = true; | 
|  | } | 
|  | } else | 
|  | pip = NULL; | 
|  | if (uip) { | 
|  | out->qs_uquota.qfs_ino = mp->m_sb.sb_uquotino; | 
|  | out->qs_uquota.qfs_nblks = uip->i_d.di_nblocks; | 
|  | out->qs_uquota.qfs_nextents = uip->i_d.di_nextents; | 
|  | if (tempuqip) | 
|  | IRELE(uip); | 
|  | } | 
|  |  | 
|  | if (gip) { | 
|  | out->qs_gquota.qfs_ino = mp->m_sb.sb_gquotino; | 
|  | out->qs_gquota.qfs_nblks = gip->i_d.di_nblocks; | 
|  | out->qs_gquota.qfs_nextents = gip->i_d.di_nextents; | 
|  | if (tempgqip) | 
|  | IRELE(gip); | 
|  | } | 
|  | if (pip) { | 
|  | out->qs_gquota.qfs_ino = mp->m_sb.sb_gquotino; | 
|  | out->qs_gquota.qfs_nblks = pip->i_d.di_nblocks; | 
|  | out->qs_gquota.qfs_nextents = pip->i_d.di_nextents; | 
|  | if (temppqip) | 
|  | IRELE(pip); | 
|  | } | 
|  | out->qs_incoredqs = q->qi_dquots; | 
|  | out->qs_btimelimit = q->qi_btimelimit; | 
|  | out->qs_itimelimit = q->qi_itimelimit; | 
|  | out->qs_rtbtimelimit = q->qi_rtbtimelimit; | 
|  | out->qs_bwarnlimit = q->qi_bwarnlimit; | 
|  | out->qs_iwarnlimit = q->qi_iwarnlimit; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Return quota status information, such as uquota-off, enforcements, etc. | 
|  | * for Q_XGETQSTATV command, to support separate project quota field. | 
|  | */ | 
|  | int | 
|  | xfs_qm_scall_getqstatv( | 
|  | struct xfs_mount	*mp, | 
|  | struct fs_quota_statv	*out) | 
|  | { | 
|  | struct xfs_quotainfo	*q = mp->m_quotainfo; | 
|  | struct xfs_inode	*uip = NULL; | 
|  | struct xfs_inode	*gip = NULL; | 
|  | struct xfs_inode	*pip = NULL; | 
|  | bool                    tempuqip = false; | 
|  | bool                    tempgqip = false; | 
|  | bool                    temppqip = false; | 
|  |  | 
|  | out->qs_flags = (__uint16_t) xfs_qm_export_flags(mp->m_qflags & | 
|  | (XFS_ALL_QUOTA_ACCT| | 
|  | XFS_ALL_QUOTA_ENFD)); | 
|  | out->qs_uquota.qfs_ino = mp->m_sb.sb_uquotino; | 
|  | out->qs_gquota.qfs_ino = mp->m_sb.sb_gquotino; | 
|  | out->qs_pquota.qfs_ino = mp->m_sb.sb_pquotino; | 
|  |  | 
|  | uip = q->qi_uquotaip; | 
|  | gip = q->qi_gquotaip; | 
|  | pip = q->qi_pquotaip; | 
|  | if (!uip && mp->m_sb.sb_uquotino != NULLFSINO) { | 
|  | if (xfs_iget(mp, NULL, mp->m_sb.sb_uquotino, | 
|  | 0, 0, &uip) == 0) | 
|  | tempuqip = true; | 
|  | } | 
|  | if (!gip && mp->m_sb.sb_gquotino != NULLFSINO) { | 
|  | if (xfs_iget(mp, NULL, mp->m_sb.sb_gquotino, | 
|  | 0, 0, &gip) == 0) | 
|  | tempgqip = true; | 
|  | } | 
|  | if (!pip && mp->m_sb.sb_pquotino != NULLFSINO) { | 
|  | if (xfs_iget(mp, NULL, mp->m_sb.sb_pquotino, | 
|  | 0, 0, &pip) == 0) | 
|  | temppqip = true; | 
|  | } | 
|  | if (uip) { | 
|  | out->qs_uquota.qfs_nblks = uip->i_d.di_nblocks; | 
|  | out->qs_uquota.qfs_nextents = uip->i_d.di_nextents; | 
|  | if (tempuqip) | 
|  | IRELE(uip); | 
|  | } | 
|  |  | 
|  | if (gip) { | 
|  | out->qs_gquota.qfs_nblks = gip->i_d.di_nblocks; | 
|  | out->qs_gquota.qfs_nextents = gip->i_d.di_nextents; | 
|  | if (tempgqip) | 
|  | IRELE(gip); | 
|  | } | 
|  | if (pip) { | 
|  | out->qs_pquota.qfs_nblks = pip->i_d.di_nblocks; | 
|  | out->qs_pquota.qfs_nextents = pip->i_d.di_nextents; | 
|  | if (temppqip) | 
|  | IRELE(pip); | 
|  | } | 
|  | out->qs_incoredqs = q->qi_dquots; | 
|  | out->qs_btimelimit = q->qi_btimelimit; | 
|  | out->qs_itimelimit = q->qi_itimelimit; | 
|  | out->qs_rtbtimelimit = q->qi_rtbtimelimit; | 
|  | out->qs_bwarnlimit = q->qi_bwarnlimit; | 
|  | out->qs_iwarnlimit = q->qi_iwarnlimit; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | #define XFS_QC_MASK \ | 
|  | (QC_LIMIT_MASK | QC_TIMER_MASK | QC_WARNS_MASK) | 
|  |  | 
|  | /* | 
|  | * Adjust quota limits, and start/stop timers accordingly. | 
|  | */ | 
|  | int | 
|  | xfs_qm_scall_setqlim( | 
|  | struct xfs_mount	*mp, | 
|  | xfs_dqid_t		id, | 
|  | uint			type, | 
|  | struct qc_dqblk		*newlim) | 
|  | { | 
|  | struct xfs_quotainfo	*q = mp->m_quotainfo; | 
|  | struct xfs_disk_dquot	*ddq; | 
|  | struct xfs_dquot	*dqp; | 
|  | struct xfs_trans	*tp; | 
|  | int			error; | 
|  | xfs_qcnt_t		hard, soft; | 
|  |  | 
|  | if (newlim->d_fieldmask & ~XFS_QC_MASK) | 
|  | return -EINVAL; | 
|  | if ((newlim->d_fieldmask & XFS_QC_MASK) == 0) | 
|  | return 0; | 
|  |  | 
|  | /* | 
|  | * We don't want to race with a quotaoff so take the quotaoff lock. | 
|  | * We don't hold an inode lock, so there's nothing else to stop | 
|  | * a quotaoff from happening. | 
|  | */ | 
|  | mutex_lock(&q->qi_quotaofflock); | 
|  |  | 
|  | /* | 
|  | * Get the dquot (locked) before we start, as we need to do a | 
|  | * transaction to allocate it if it doesn't exist. Once we have the | 
|  | * dquot, unlock it so we can start the next transaction safely. We hold | 
|  | * a reference to the dquot, so it's safe to do this unlock/lock without | 
|  | * it being reclaimed in the mean time. | 
|  | */ | 
|  | error = xfs_qm_dqget(mp, NULL, id, type, XFS_QMOPT_DQALLOC, &dqp); | 
|  | if (error) { | 
|  | ASSERT(error != -ENOENT); | 
|  | goto out_unlock; | 
|  | } | 
|  | xfs_dqunlock(dqp); | 
|  |  | 
|  | tp = xfs_trans_alloc(mp, XFS_TRANS_QM_SETQLIM); | 
|  | error = xfs_trans_reserve(tp, &M_RES(mp)->tr_qm_setqlim, 0, 0); | 
|  | if (error) { | 
|  | xfs_trans_cancel(tp, 0); | 
|  | goto out_rele; | 
|  | } | 
|  |  | 
|  | xfs_dqlock(dqp); | 
|  | xfs_trans_dqjoin(tp, dqp); | 
|  | ddq = &dqp->q_core; | 
|  |  | 
|  | /* | 
|  | * Make sure that hardlimits are >= soft limits before changing. | 
|  | */ | 
|  | hard = (newlim->d_fieldmask & QC_SPC_HARD) ? | 
|  | (xfs_qcnt_t) XFS_B_TO_FSB(mp, newlim->d_spc_hardlimit) : | 
|  | be64_to_cpu(ddq->d_blk_hardlimit); | 
|  | soft = (newlim->d_fieldmask & QC_SPC_SOFT) ? | 
|  | (xfs_qcnt_t) XFS_B_TO_FSB(mp, newlim->d_spc_softlimit) : | 
|  | be64_to_cpu(ddq->d_blk_softlimit); | 
|  | if (hard == 0 || hard >= soft) { | 
|  | ddq->d_blk_hardlimit = cpu_to_be64(hard); | 
|  | ddq->d_blk_softlimit = cpu_to_be64(soft); | 
|  | xfs_dquot_set_prealloc_limits(dqp); | 
|  | if (id == 0) { | 
|  | q->qi_bhardlimit = hard; | 
|  | q->qi_bsoftlimit = soft; | 
|  | } | 
|  | } else { | 
|  | xfs_debug(mp, "blkhard %Ld < blksoft %Ld", hard, soft); | 
|  | } | 
|  | hard = (newlim->d_fieldmask & QC_RT_SPC_HARD) ? | 
|  | (xfs_qcnt_t) XFS_B_TO_FSB(mp, newlim->d_rt_spc_hardlimit) : | 
|  | be64_to_cpu(ddq->d_rtb_hardlimit); | 
|  | soft = (newlim->d_fieldmask & QC_RT_SPC_SOFT) ? | 
|  | (xfs_qcnt_t) XFS_B_TO_FSB(mp, newlim->d_rt_spc_softlimit) : | 
|  | be64_to_cpu(ddq->d_rtb_softlimit); | 
|  | if (hard == 0 || hard >= soft) { | 
|  | ddq->d_rtb_hardlimit = cpu_to_be64(hard); | 
|  | ddq->d_rtb_softlimit = cpu_to_be64(soft); | 
|  | if (id == 0) { | 
|  | q->qi_rtbhardlimit = hard; | 
|  | q->qi_rtbsoftlimit = soft; | 
|  | } | 
|  | } else { | 
|  | xfs_debug(mp, "rtbhard %Ld < rtbsoft %Ld", hard, soft); | 
|  | } | 
|  |  | 
|  | hard = (newlim->d_fieldmask & QC_INO_HARD) ? | 
|  | (xfs_qcnt_t) newlim->d_ino_hardlimit : | 
|  | be64_to_cpu(ddq->d_ino_hardlimit); | 
|  | soft = (newlim->d_fieldmask & QC_INO_SOFT) ? | 
|  | (xfs_qcnt_t) newlim->d_ino_softlimit : | 
|  | be64_to_cpu(ddq->d_ino_softlimit); | 
|  | if (hard == 0 || hard >= soft) { | 
|  | ddq->d_ino_hardlimit = cpu_to_be64(hard); | 
|  | ddq->d_ino_softlimit = cpu_to_be64(soft); | 
|  | if (id == 0) { | 
|  | q->qi_ihardlimit = hard; | 
|  | q->qi_isoftlimit = soft; | 
|  | } | 
|  | } else { | 
|  | xfs_debug(mp, "ihard %Ld < isoft %Ld", hard, soft); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Update warnings counter(s) if requested | 
|  | */ | 
|  | if (newlim->d_fieldmask & QC_SPC_WARNS) | 
|  | ddq->d_bwarns = cpu_to_be16(newlim->d_spc_warns); | 
|  | if (newlim->d_fieldmask & QC_INO_WARNS) | 
|  | ddq->d_iwarns = cpu_to_be16(newlim->d_ino_warns); | 
|  | if (newlim->d_fieldmask & QC_RT_SPC_WARNS) | 
|  | ddq->d_rtbwarns = cpu_to_be16(newlim->d_rt_spc_warns); | 
|  |  | 
|  | if (id == 0) { | 
|  | /* | 
|  | * Timelimits for the super user set the relative time | 
|  | * the other users can be over quota for this file system. | 
|  | * If it is zero a default is used.  Ditto for the default | 
|  | * soft and hard limit values (already done, above), and | 
|  | * for warnings. | 
|  | */ | 
|  | if (newlim->d_fieldmask & QC_SPC_TIMER) { | 
|  | q->qi_btimelimit = newlim->d_spc_timer; | 
|  | ddq->d_btimer = cpu_to_be32(newlim->d_spc_timer); | 
|  | } | 
|  | if (newlim->d_fieldmask & QC_INO_TIMER) { | 
|  | q->qi_itimelimit = newlim->d_ino_timer; | 
|  | ddq->d_itimer = cpu_to_be32(newlim->d_ino_timer); | 
|  | } | 
|  | if (newlim->d_fieldmask & QC_RT_SPC_TIMER) { | 
|  | q->qi_rtbtimelimit = newlim->d_rt_spc_timer; | 
|  | ddq->d_rtbtimer = cpu_to_be32(newlim->d_rt_spc_timer); | 
|  | } | 
|  | if (newlim->d_fieldmask & QC_SPC_WARNS) | 
|  | q->qi_bwarnlimit = newlim->d_spc_warns; | 
|  | if (newlim->d_fieldmask & QC_INO_WARNS) | 
|  | q->qi_iwarnlimit = newlim->d_ino_warns; | 
|  | if (newlim->d_fieldmask & QC_RT_SPC_WARNS) | 
|  | q->qi_rtbwarnlimit = newlim->d_rt_spc_warns; | 
|  | } else { | 
|  | /* | 
|  | * If the user is now over quota, start the timelimit. | 
|  | * The user will not be 'warned'. | 
|  | * Note that we keep the timers ticking, whether enforcement | 
|  | * is on or off. We don't really want to bother with iterating | 
|  | * over all ondisk dquots and turning the timers on/off. | 
|  | */ | 
|  | xfs_qm_adjust_dqtimers(mp, ddq); | 
|  | } | 
|  | dqp->dq_flags |= XFS_DQ_DIRTY; | 
|  | xfs_trans_log_dquot(tp, dqp); | 
|  |  | 
|  | error = xfs_trans_commit(tp, 0); | 
|  |  | 
|  | out_rele: | 
|  | xfs_qm_dqrele(dqp); | 
|  | out_unlock: | 
|  | mutex_unlock(&q->qi_quotaofflock); | 
|  | return error; | 
|  | } | 
|  |  | 
|  | STATIC int | 
|  | xfs_qm_log_quotaoff_end( | 
|  | xfs_mount_t		*mp, | 
|  | xfs_qoff_logitem_t	*startqoff, | 
|  | uint			flags) | 
|  | { | 
|  | xfs_trans_t		*tp; | 
|  | int			error; | 
|  | xfs_qoff_logitem_t	*qoffi; | 
|  |  | 
|  | tp = xfs_trans_alloc(mp, XFS_TRANS_QM_QUOTAOFF_END); | 
|  |  | 
|  | error = xfs_trans_reserve(tp, &M_RES(mp)->tr_qm_equotaoff, 0, 0); | 
|  | if (error) { | 
|  | xfs_trans_cancel(tp, 0); | 
|  | return error; | 
|  | } | 
|  |  | 
|  | qoffi = xfs_trans_get_qoff_item(tp, startqoff, | 
|  | flags & XFS_ALL_QUOTA_ACCT); | 
|  | xfs_trans_log_quotaoff_item(tp, qoffi); | 
|  |  | 
|  | /* | 
|  | * We have to make sure that the transaction is secure on disk before we | 
|  | * return and actually stop quota accounting. So, make it synchronous. | 
|  | * We don't care about quotoff's performance. | 
|  | */ | 
|  | xfs_trans_set_sync(tp); | 
|  | error = xfs_trans_commit(tp, 0); | 
|  | return error; | 
|  | } | 
|  |  | 
|  |  | 
|  | STATIC int | 
|  | xfs_qm_log_quotaoff( | 
|  | xfs_mount_t	       *mp, | 
|  | xfs_qoff_logitem_t     **qoffstartp, | 
|  | uint		       flags) | 
|  | { | 
|  | xfs_trans_t	       *tp; | 
|  | int			error; | 
|  | xfs_qoff_logitem_t     *qoffi; | 
|  |  | 
|  | *qoffstartp = NULL; | 
|  |  | 
|  | tp = xfs_trans_alloc(mp, XFS_TRANS_QM_QUOTAOFF); | 
|  | error = xfs_trans_reserve(tp, &M_RES(mp)->tr_qm_quotaoff, 0, 0); | 
|  | if (error) { | 
|  | xfs_trans_cancel(tp, 0); | 
|  | goto out; | 
|  | } | 
|  |  | 
|  | qoffi = xfs_trans_get_qoff_item(tp, NULL, flags & XFS_ALL_QUOTA_ACCT); | 
|  | xfs_trans_log_quotaoff_item(tp, qoffi); | 
|  |  | 
|  | spin_lock(&mp->m_sb_lock); | 
|  | mp->m_sb.sb_qflags = (mp->m_qflags & ~(flags)) & XFS_MOUNT_QUOTA_ALL; | 
|  | spin_unlock(&mp->m_sb_lock); | 
|  |  | 
|  | xfs_log_sb(tp); | 
|  |  | 
|  | /* | 
|  | * We have to make sure that the transaction is secure on disk before we | 
|  | * return and actually stop quota accounting. So, make it synchronous. | 
|  | * We don't care about quotoff's performance. | 
|  | */ | 
|  | xfs_trans_set_sync(tp); | 
|  | error = xfs_trans_commit(tp, 0); | 
|  | if (error) | 
|  | goto out; | 
|  |  | 
|  | *qoffstartp = qoffi; | 
|  | out: | 
|  | return error; | 
|  | } | 
|  |  | 
|  |  | 
|  | int | 
|  | xfs_qm_scall_getquota( | 
|  | struct xfs_mount	*mp, | 
|  | xfs_dqid_t		id, | 
|  | uint			type, | 
|  | struct qc_dqblk		*dst) | 
|  | { | 
|  | struct xfs_dquot	*dqp; | 
|  | int			error; | 
|  |  | 
|  | /* | 
|  | * Try to get the dquot. We don't want it allocated on disk, so | 
|  | * we aren't passing the XFS_QMOPT_DOALLOC flag. If it doesn't | 
|  | * exist, we'll get ENOENT back. | 
|  | */ | 
|  | error = xfs_qm_dqget(mp, NULL, id, type, 0, &dqp); | 
|  | if (error) | 
|  | return error; | 
|  |  | 
|  | /* | 
|  | * If everything's NULL, this dquot doesn't quite exist as far as | 
|  | * our utility programs are concerned. | 
|  | */ | 
|  | if (XFS_IS_DQUOT_UNINITIALIZED(dqp)) { | 
|  | error = -ENOENT; | 
|  | goto out_put; | 
|  | } | 
|  |  | 
|  | memset(dst, 0, sizeof(*dst)); | 
|  | dst->d_spc_hardlimit = | 
|  | XFS_FSB_TO_B(mp, be64_to_cpu(dqp->q_core.d_blk_hardlimit)); | 
|  | dst->d_spc_softlimit = | 
|  | XFS_FSB_TO_B(mp, be64_to_cpu(dqp->q_core.d_blk_softlimit)); | 
|  | dst->d_ino_hardlimit = be64_to_cpu(dqp->q_core.d_ino_hardlimit); | 
|  | dst->d_ino_softlimit = be64_to_cpu(dqp->q_core.d_ino_softlimit); | 
|  | dst->d_space = XFS_FSB_TO_B(mp, dqp->q_res_bcount); | 
|  | dst->d_ino_count = dqp->q_res_icount; | 
|  | dst->d_spc_timer = be32_to_cpu(dqp->q_core.d_btimer); | 
|  | dst->d_ino_timer = be32_to_cpu(dqp->q_core.d_itimer); | 
|  | dst->d_ino_warns = be16_to_cpu(dqp->q_core.d_iwarns); | 
|  | dst->d_spc_warns = be16_to_cpu(dqp->q_core.d_bwarns); | 
|  | dst->d_rt_spc_hardlimit = | 
|  | XFS_FSB_TO_B(mp, be64_to_cpu(dqp->q_core.d_rtb_hardlimit)); | 
|  | dst->d_rt_spc_softlimit = | 
|  | XFS_FSB_TO_B(mp, be64_to_cpu(dqp->q_core.d_rtb_softlimit)); | 
|  | dst->d_rt_space = XFS_FSB_TO_B(mp, dqp->q_res_rtbcount); | 
|  | dst->d_rt_spc_timer = be32_to_cpu(dqp->q_core.d_rtbtimer); | 
|  | dst->d_rt_spc_warns = be16_to_cpu(dqp->q_core.d_rtbwarns); | 
|  |  | 
|  | /* | 
|  | * Internally, we don't reset all the timers when quota enforcement | 
|  | * gets turned off. No need to confuse the user level code, | 
|  | * so return zeroes in that case. | 
|  | */ | 
|  | if ((!XFS_IS_UQUOTA_ENFORCED(mp) && | 
|  | dqp->q_core.d_flags == XFS_DQ_USER) || | 
|  | (!XFS_IS_GQUOTA_ENFORCED(mp) && | 
|  | dqp->q_core.d_flags == XFS_DQ_GROUP) || | 
|  | (!XFS_IS_PQUOTA_ENFORCED(mp) && | 
|  | dqp->q_core.d_flags == XFS_DQ_PROJ)) { | 
|  | dst->d_spc_timer = 0; | 
|  | dst->d_ino_timer = 0; | 
|  | dst->d_rt_spc_timer = 0; | 
|  | } | 
|  |  | 
|  | #ifdef DEBUG | 
|  | if (((XFS_IS_UQUOTA_ENFORCED(mp) && type == XFS_DQ_USER) || | 
|  | (XFS_IS_GQUOTA_ENFORCED(mp) && type == XFS_DQ_GROUP) || | 
|  | (XFS_IS_PQUOTA_ENFORCED(mp) && type == XFS_DQ_PROJ)) && | 
|  | id != 0) { | 
|  | if ((dst->d_space > dst->d_spc_softlimit) && | 
|  | (dst->d_spc_softlimit > 0)) { | 
|  | ASSERT(dst->d_spc_timer != 0); | 
|  | } | 
|  | if ((dst->d_ino_count > dst->d_ino_softlimit) && | 
|  | (dst->d_ino_softlimit > 0)) { | 
|  | ASSERT(dst->d_ino_timer != 0); | 
|  | } | 
|  | } | 
|  | #endif | 
|  | out_put: | 
|  | xfs_qm_dqput(dqp); | 
|  | return error; | 
|  | } | 
|  |  | 
|  | STATIC uint | 
|  | xfs_qm_export_flags( | 
|  | uint flags) | 
|  | { | 
|  | uint uflags; | 
|  |  | 
|  | uflags = 0; | 
|  | if (flags & XFS_UQUOTA_ACCT) | 
|  | uflags |= FS_QUOTA_UDQ_ACCT; | 
|  | if (flags & XFS_GQUOTA_ACCT) | 
|  | uflags |= FS_QUOTA_GDQ_ACCT; | 
|  | if (flags & XFS_PQUOTA_ACCT) | 
|  | uflags |= FS_QUOTA_PDQ_ACCT; | 
|  | if (flags & XFS_UQUOTA_ENFD) | 
|  | uflags |= FS_QUOTA_UDQ_ENFD; | 
|  | if (flags & XFS_GQUOTA_ENFD) | 
|  | uflags |= FS_QUOTA_GDQ_ENFD; | 
|  | if (flags & XFS_PQUOTA_ENFD) | 
|  | uflags |= FS_QUOTA_PDQ_ENFD; | 
|  | return uflags; | 
|  | } | 
|  |  | 
|  |  | 
|  | STATIC int | 
|  | xfs_dqrele_inode( | 
|  | struct xfs_inode	*ip, | 
|  | int			flags, | 
|  | void			*args) | 
|  | { | 
|  | /* skip quota inodes */ | 
|  | if (ip == ip->i_mount->m_quotainfo->qi_uquotaip || | 
|  | ip == ip->i_mount->m_quotainfo->qi_gquotaip || | 
|  | ip == ip->i_mount->m_quotainfo->qi_pquotaip) { | 
|  | ASSERT(ip->i_udquot == NULL); | 
|  | ASSERT(ip->i_gdquot == NULL); | 
|  | ASSERT(ip->i_pdquot == NULL); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | xfs_ilock(ip, XFS_ILOCK_EXCL); | 
|  | if ((flags & XFS_UQUOTA_ACCT) && ip->i_udquot) { | 
|  | xfs_qm_dqrele(ip->i_udquot); | 
|  | ip->i_udquot = NULL; | 
|  | } | 
|  | if ((flags & XFS_GQUOTA_ACCT) && ip->i_gdquot) { | 
|  | xfs_qm_dqrele(ip->i_gdquot); | 
|  | ip->i_gdquot = NULL; | 
|  | } | 
|  | if ((flags & XFS_PQUOTA_ACCT) && ip->i_pdquot) { | 
|  | xfs_qm_dqrele(ip->i_pdquot); | 
|  | ip->i_pdquot = NULL; | 
|  | } | 
|  | xfs_iunlock(ip, XFS_ILOCK_EXCL); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  |  | 
|  | /* | 
|  | * Go thru all the inodes in the file system, releasing their dquots. | 
|  | * | 
|  | * Note that the mount structure gets modified to indicate that quotas are off | 
|  | * AFTER this, in the case of quotaoff. | 
|  | */ | 
|  | void | 
|  | xfs_qm_dqrele_all_inodes( | 
|  | struct xfs_mount *mp, | 
|  | uint		 flags) | 
|  | { | 
|  | ASSERT(mp->m_quotainfo); | 
|  | xfs_inode_ag_iterator(mp, xfs_dqrele_inode, flags, NULL); | 
|  | } |